Modelforge is a CVEDIA proprietary API to make AI models available on any platform without any extra heavy lifting by the user of CVEDIA-RT. Simply choose the desired AI model to load, and Modelforge will make it available for your platforms.
listnndevices to find available devices¶
CVEDIA-RT package comes with a standalone application called listnndevices. When running this application you will get a list of all the available devices on the machine.
modelforge to query available models for each device¶
CVEDIA-RT package comes with a standalone application called modelforge. When running this application you will get a list of all the models that can be automatically downloaded for the machine's device.
The query must be made specifying what device to use, based on the available devices listed on the listnndevices application.
modelforge.exe --uri [DEVICE]
modelforge.exe --uri tensorrt.1://
Seeing verbose output
If the -v flag is used when running the modelforge command, a verbose debug output will be presented
Using the model on the application¶
After choosing the model from the list returned by the Modelforge application, you can set it on the project settings JSON file in the corresponding inference section:
auto: on the device section of the json
If the user specifies
auto: as the device for the model to load, CVEDIA-RT will automatically choose the better device for that machine to run the model.
When the instance starts, CVEDIA-RT will automatically reach the Model Forge servers and download the model corresponding to the configured settings, which will then use for inference.
Using a local model instead of CVEDIA's models¶
You can use your own models instead of ones delivered by our system.
/. This will instruct CVEDIA-RT to load the file from a local path.
A model should have a
.json file with its configuration parameters:
"comment": "Model name",
You can find more on how to configure it in inference