GpuAccelerationConfig

  • GpuAccelerationConfig is a concrete class representing GPU acceleration configurations and extends AccelerationConfig.

  • It includes nested classes for building configurations, selecting the GPU backend, and defining inference priorities and usage preferences.

  • The class provides public methods to retrieve details such as the cache directory, quantized inference flag, force backend setting, inference preference, and multiple inference priority levels.

  • It also offers a method to retrieve a unique model token string.

  • Inherited methods from AccelerationConfig allow retrieving the accelerator name and serializing the configuration.

public class GpuAccelerationConfig extends AccelerationConfig

Concrete class that represents GPU acceleration configs. For more details, see delegate.h

Nested Class Summary

class GpuAccelerationConfig.Builder Builder class. 
enum GpuAccelerationConfig.GpuBackend Which GPU backend to select. 
enum GpuAccelerationConfig.GpuInferencePriority Relative priorities given by the GPU delegate to different client needs. 
enum GpuAccelerationConfig.GpuInferenceUsage GPU inference preference for initialization time vs. 

Public Method Summary

String
cacheDirectory()
Returns serialization cache directory.
boolean
enableQuantizedInference()
Returns the enable quantized inference flag.
GpuAccelerationConfig.GpuBackend
forceBackend()
Returns the selected GPU backend.
GpuAccelerationConfig.GpuInferenceUsage
inferencePreference()
Returns GPU inference preference.
GpuAccelerationConfig.GpuInferencePriority
inferencePriority1()
Returns GPU inference priority(1).
GpuAccelerationConfig.GpuInferencePriority
inferencePriority2()
Returns GPU inference priority(2).
GpuAccelerationConfig.GpuInferencePriority
inferencePriority3()
Returns GPU inference priority(3).
String
modelToken()
Returns unique model token string.

Inherited Method Summary

Public Methods

public String cacheDirectory ()

Returns serialization cache directory.

public boolean enableQuantizedInference ()

Returns the enable quantized inference flag.

public GpuAccelerationConfig.GpuBackend forceBackend ()

Returns the selected GPU backend.

public GpuAccelerationConfig.GpuInferenceUsage inferencePreference ()

Returns GPU inference preference.

public GpuAccelerationConfig.GpuInferencePriority inferencePriority1 ()

Returns GPU inference priority(1).

public GpuAccelerationConfig.GpuInferencePriority inferencePriority2 ()

Returns GPU inference priority(2).

public GpuAccelerationConfig.GpuInferencePriority inferencePriority3 ()

Returns GPU inference priority(3).

public String modelToken ()

Returns unique model token string.