-
Notifications
You must be signed in to change notification settings - Fork 241
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pre-fetching jars in docker environment fails to populate classpath #1265
Comments
Almond uses a separate directory for cache. |
That doesn't appear to be the case. Both methods (import from notebook and import $ivy.`org.apache.logging.log4j:log4j-core:2.17.0` are
I'm thinking that the way to pre-load is to provide a notebook with the desired inputs and run it through jupyter during the docker build. There appears to be some amount of state that is created in in-notebook imports that Edit: |
hmm I did not observe this with the docker image I'm using. However, I'm using
in the dockerfile. Does that impact the coursier cache for even the notebook session? https://github.com/coreyoconnor/nix_configs/blob/dev/modules/ufo-k8s/almond-2/Dockerfile |
After further testing. Yes, setting |
I want to build a docker environment where I can pre-load the classpath with
spark-sql
and some other stuff to avoid boilerplate in my notebooks. So I built the following Dockerfile:However, upon running this container, running
import org.apache.spark.sql._
yields an error:What step am I missing to get Almond to recognize the coursier-installed jars?
The text was updated successfully, but these errors were encountered: