Skip to content

Fixed ml.get_memory_stats after tests #1578

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Mar 28, 2022
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
51 changes: 35 additions & 16 deletions output/schema/schema.json

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

25 changes: 13 additions & 12 deletions output/typescript/types.ts

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ import { Memory } from './types'
export class Response {
body: {
_nodes: NodeStatistics
cluser_name: Name
cluster_name: Name
nodes: Dictionary<Id, Memory>
}
}
26 changes: 14 additions & 12 deletions specification/ml/get_memory_stats/types.ts
Original file line number Diff line number Diff line change
Expand Up @@ -17,12 +17,13 @@
* under the License.
*/

import { ByteSize, Name } from '@_types/common'
import { Dictionary } from '@spec_utils/Dictionary'
import { ByteSize, Id, Name } from '@_types/common'
import { TransportAddress } from '@_types/Networking'
import { integer } from '@_types/Numeric'

export class Memory {
attributes: string[]
attributes: Dictionary<string, string>
/**
* Contains Java Virtual Machine (JVM) statistics for the node.
*/
Expand All @@ -43,19 +44,20 @@ export class Memory {
* The host and port where transport HTTP connections are accepted.
*/
transport_address: TransportAddress
ephemeral_id: Id
}

export class JvmStats {
/** Maximum amount of memory available for use by the heap. */
heap_max: ByteSize
heap_max?: ByteSize
/** Maximum amount of memory, in bytes, available for use by the heap. */
heap_max_in_bytes: integer
/** Amount of Java heap currently being used for caching inference models. */
java_inference: ByteSize
java_inference?: ByteSize
/** Amount of Java heap, in bytes, currently being used for caching inference models. */
java_inference_in_bytes: integer
/** Maximum amount of Java heap to be used for caching inference models. */
java_inference_max: ByteSize
java_inference_max?: ByteSize
/** Maximum amount of Java heap, in bytes, to be used for caching inference models. */
java_inference_max_in_bytes: integer
}
Expand All @@ -65,7 +67,7 @@ export class MemStats {
* If the amount of physical memory has been overridden using the es.total_memory_bytes system property
* then this reports the overridden value. Otherwise it reports the same value as total.
*/
adjusted_total: ByteSize
adjusted_total?: ByteSize
/**
* If the amount of physical memory has been overridden using the `es.total_memory_bytes` system property
* then this reports the overridden value in bytes. Otherwise it reports the same value as `total_in_bytes`.
Expand All @@ -74,7 +76,7 @@ export class MemStats {
/**
* Total amount of physical memory.
*/
total: ByteSize
total?: ByteSize
/**
* Total amount of physical memory in bytes.
*/
Expand All @@ -87,23 +89,23 @@ export class MemStats {

export class MemMlStats {
/** Amount of native memory set aside for anomaly detection jobs. */
anomaly_detectors: ByteSize
anomaly_detectors?: ByteSize
/** Amount of native memory, in bytes, set aside for anomaly detection jobs. */
anomaly_detectors_in_bytes: integer
/** Amount of native memory set aside for data frame analytics jobs. */
data_frame_analytics: ByteSize
data_frame_analytics?: ByteSize
/** Amount of native memory, in bytes, set aside for data frame analytics jobs. */
data_frame_analytics_in_bytes: integer
/** Maximum amount of native memory (separate to the JVM heap) that may be used by machine learning native processes. */
max: ByteSize
max?: ByteSize
/** Maximum amount of native memory (separate to the JVM heap), in bytes, that may be used by machine learning native processes. */
max_in_bytes: integer
/** Amount of native memory set aside for loading machine learning native code shared libraries. */
native_code_overhead: ByteSize
native_code_overhead?: ByteSize
/** Amount of native memory, in bytes, set aside for loading machine learning native code shared libraries. */
native_code_overhead_in_bytes: integer
/** Amount of native memory set aside for trained models that have a PyTorch model_type. */
native_inference: ByteSize
native_inference?: ByteSize
/** Amount of native memory, in bytes, set aside for trained models that have a PyTorch model_type. */
native_inference_in_bytes: integer
}