You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello,
I can't get the babar.log aggregated to yarn logs. When I run the command to get the logs: yarn logs --applicationId application_XXXX_YYYY > myAppLog.log
the resulting myAppLog.log doesn't contain the traces of babar.log. It contains only my application and yarn log messages.
It is not a problem when I launch the application from the linux command line (calling directly spark2-submit) because the file babar.log is created in the same directory. But when I launch it from an oozie workflow (production environment), the file babar.log dissapears when the container terminates and its content is not aggregated.
I realized that the system properties: yarn.app.container.log.dir and spark.yarn.app.container.log.dir are null, then babar uses a local directory where log is stored : ./log. Could it be the reason? Anyone has observed the same problem?
The text was updated successfully, but these errors were encountered:
Hello!
Yes it could be the problem, the agent uses the environment properties to find where to store the log file. It the properties are not found, it will store to a local folder named log.
If you tried to replace the environemnt property by getting the value from an initialized hadoop configuration, would you get a correctly set value?
Something like:
Configuration conf = new Configuration();
conf.get("yarn.app.container.log.dir")
You could also try to specify a custom log directory from the agent parameters that you add to your java options.
Hi Benoit,
thanks for your support. We've chosen the easiest solution, write the babar.log to a local directory outside the container (/tmp) and collect it afterwards manually from the concerned server (driver's server in our case). I couldn't find an easy solution to get the property value (directory of yarn containers' logs) and pass it to babar before launching spark2-submit from an oozie workflow.
We activated babar only for the driver, we didn't arrive to launch it in the executors. But as we're actually working in a problem with the driver, then it's enough by now.
Hello,
I can't get the babar.log aggregated to yarn logs. When I run the command to get the logs:
yarn logs --applicationId application_XXXX_YYYY > myAppLog.log
the resulting
myAppLog.log
doesn't contain the traces ofbabar.log
. It contains only my application and yarn log messages.It is not a problem when I launch the application from the linux command line (calling directly spark2-submit) because the file
babar.log
is created in the same directory. But when I launch it from an oozie workflow (production environment), the filebabar.log
dissapears when the container terminates and its content is not aggregated.I realized that the system properties:
yarn.app.container.log.dir
andspark.yarn.app.container.log.dir
arenull
, thenbabar
uses a local directory where log is stored :./log
. Could it be the reason? Anyone has observed the same problem?The text was updated successfully, but these errors were encountered: