Hi @alfarzan,
My latest submissions (e.g. 114780) have encountered a problem with the same error message, which this thread is reffering to: “Inference failed. View the submission for more details.”.
In the log I can find following detailed error message:
“Error in load_model(model_output_path) :
unused argument (model_output_path)”
I really can’t find any solution. Can you please take a look at my submission?
Can you please provide a submission number so I can look into this one? I’m looking into the one from @Baracuda already as they’ve reached out on Discord as well
@actuarist for @Baracuda who is using zip submissions the issue was that the save and load model functions need to be altered to work with how h2o models are saved and loaded. Let me know if this doesn’t resolve your issue as well.
finally working
been mostly using the colab colab seems to run fine but can’t seem to get submission working
worked with zip submission after adding in initialising library statements in predict.R
i thought it should automatically initialise packages in the install.R?
In colab any initialisation you need should be done inside of global_imports which is sourced at the top of every file including in predict.R. In zip you are to do that inside of model.R which gets sourced at the top of predict.R.
Can I have your submission ID to make sure things are working fine from our end?
For notebook submissions the only difference would be that the the hard-coded path of the saved model has to be in a directory called saved_objects. Otherwise it should work very similarly with the exception of having to use the install_packages and global_imports functions. h2o.init() would fit into the latter.
Hi, I’ve been having an issue with a Lightgbm model. The training works but I get an error importing lightgbm during inference.
########## Generating Predictions on /data/mse_debug_data.csv #############
Traceback (most recent call last):
File “predict.py”, line 20, in
from utils import *
File “/home/aicrowd/utils.py”, line 1, in
from global_imports import *
File “/home/aicrowd/global_imports.py”, line 5, in
import lightgbm as lgb
File “/usr/local/lib/python3.8/site-packages/lightgbm/init.py”, line 8, in
from .basic import Booster, Dataset
File “/usr/local/lib/python3.8/site-packages/lightgbm/basic.py”, line 43, in
_LIB = _load_lib()
File “/usr/local/lib/python3.8/site-packages/lightgbm/basic.py”, line 34, in _load_lib
lib = ctypes.cdll.LoadLibrary(lib_path[0])
File “/usr/local/lib/python3.8/ctypes/init.py”, line 451, in LoadLibrary
return self._dlltype(name)
File “/usr/local/lib/python3.8/ctypes/init.py”, line 373, in init
self._handle = _dlopen(self._name, mode)
OSError: libgomp.so.1: cannot open shared object file: No such file or directory
I keep getting Inference error messages, even thought my notebook codes works correctly. Can you help me figure out what is going on? Submission # 118475
Yes I see what is going on. I think you may be using one of the notebook versions from the early weeks of the game. We made a small change to the load_model function by making sure it requires no arguments.
In your case the fix is easy, can you please replace your load_model function with this?
load_model <- function(){
# Load a saved trained model from the file `trained_model.RData`.
# This is called by the server to evaluate your submission on hidden data.
# Only modify this *if* you modified save_model.
load('trained_model.RData')
return(model)
}
I can generate prices using your submission with the following change