The API supports ControlNet, but the input image has to be a normal image which later is preprocessed to be fed to ControlNet. For some use cases is useful to input an already processed image, such as a depth map, or a pose. This way the user can have more control over the image.
This feature is already supported in the AI generator, but not in the API.