-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add NIM saved views and logs pipeline #19113
Merged
Merged
Changes from 13 commits
Commits
Show all changes
15 commits
Select commit
Hold shift + click to select a range
db17922
Create Nvidia NIM scaffolding
Kyle-Neale 3b2bdf3
Add Initial Release changelog
Kyle-Neale 6281eaf
sync models and config
Kyle-Neale 076348d
Add metadata and tests
Kyle-Neale d9ca762
Add Readme
Kyle-Neale 08738a7
nvidia dash (#19074)
steveny91 af0725a
validate-assets fixes
Kyle-Neale 2059404
remove astericks in README hyperlink ref
Kyle-Neale 2d64b54
Add NIM saved views and logs pipeline
Kyle-Neale cbd2776
Fix add saved view to manifest.json
Kyle-Neale cd9b3ae
yaml fixes
Kyle-Neale f1378ef
Merge branch 'master' into kyle.neale/nvidia-nim-logs
Kyle-Neale a0654b5
Fix metric_id and log parsing
Kyle-Neale f6e37a6
use notSpace instead of data
Kyle-Neale 8047e14
Merge branch 'master' into kyle.neale/nvidia-nim-logs
Kyle-Neale File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,47 @@ | ||
id: nvidia_nim | ||
metric_id: nvidia-nim | ||
backend_only: false | ||
facets: | ||
pipeline: | ||
type: pipeline | ||
name: 'NVIDIA NIM' | ||
enabled: true | ||
filter: | ||
query: source:nvidia_nim | ||
processors: | ||
- type: grok-parser | ||
name: Parse timestamp, level, logger, and message | ||
enabled: true | ||
source: message | ||
samples: | ||
- "2024-10-30 21:56:25,295 [INFO] PyTorch version 2.3.1 available." | ||
- "2024-10-30 21:58:26,914 [WARNING] [TRT-LLM] [W] Logger level already set from environment. Discard new verbosity: error" | ||
- "INFO 2024-10-30 21:56:28.831 ngc_injector.py:152] Valid profile: e45b4b991bbc51d0df3ce53e87060fc3a7f76555406ed534a8479c6faa706987 (tensorrt_llm-a10g-bf16-tp4-latency) on GPUs [0, 1, 2, 3]" | ||
- "WARNING 2024-10-30 21:58:27.670 arg_utils.py:775] Chunked prefill is enabled by default for models with max_model_len > 32K. Currently, chunked prefill might not work with some features or models. If you encounter any issues, please disable chunked prefill by setting --enable-chunked-prefill=False." | ||
- "[1730325496.647520] [dd317ab0670e:126 :0] parser.c:2305 UCX WARN (set UCX_WARN_UNUSED_ENV_VARS=n to suppress this warning)" | ||
grok: | ||
matchRules: | | ||
nvidia_nim %{date("yyyy-MM-dd HH:mm:ss,SSS"):timestamp} \[%{_level}\] \[%{data:component_name}\] \[%{word}\] %{_msg} | ||
nvidia_nim_logger %{_level} %{date("yyyy-MM-dd HH:mm:ss.SSS"):timestamp} %{_logger_name}:%{_logger_line}\] %{_msg} | ||
generic_log %{date("yyyy-MM-dd HH:mm:ss,SSS"):timestamp} \[%{_level}\] %{_msg} | ||
componont_log \[%{number:timestamp}\]\W+\[%{data:container_id}:%{number:pid}\W+:%{number:thread_id}\W+%{_logger_name}:%{_logger_line} %{word:component_name}\W+%{_level}\W+\(%{_msg}\) | ||
supportRules: | | ||
_logger_line %{data:logger.line} | ||
_logger_name %{data:logger.name} | ||
_level %{word:level} | ||
_msg %{data:msg} | ||
- type: message-remapper | ||
name: Define `msg` as the official message of the log | ||
enabled: true | ||
sources: | ||
- msg | ||
- type: date-remapper | ||
name: Define `timestamp` as the official date of the log | ||
enabled: true | ||
sources: | ||
- timestamp | ||
- type: status-remapper | ||
name: Define `level` as the official status of the log | ||
enabled: true | ||
sources: | ||
- level |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,58 @@ | ||
id: "nvidia_nim" | ||
tests: | ||
# This log sample satisfies the validation. | ||
- | ||
sample: |- | ||
2024-10-30 21:56:25,295 [INFO] PyTorch version 2.3.1 available. | ||
result: | ||
custom: | ||
level: "INFO" | ||
timestamp: 1730325385295 | ||
message: "PyTorch version 2.3.1 available." | ||
status: "info" | ||
tags: | ||
- "source:LOGS_SOURCE" | ||
timestamp: 1730325385295 | ||
- | ||
sample: |- | ||
2024-10-30 21:58:26,914 [WARNING] [TRT-LLM] [W] Logger level already set from environment. Discard new verbosity: error | ||
result: | ||
custom: | ||
level: "WARNING" | ||
timestamp: 1730325506914 | ||
component_name: "TRT-LLM" | ||
message: "Logger level already set from environment. Discard new verbosity: error" | ||
status: "warn" | ||
tags: | ||
- "source:LOGS_SOURCE" | ||
timestamp: 1730325506914 | ||
- | ||
sample: |- | ||
INFO 2024-10-30 21:56:28.831 ngc_injector.py:152] Valid profile: e45b4b991bbc51d0df3ce53e87060fc3a7f76555406ed534a8479c6faa706987 (tensorrt_llm-a10g-bf16-tp4-latency) on GPUs [0, 1, 2, 3] | ||
result: | ||
custom: | ||
level: "INFO" | ||
timestamp: 1730325388831 | ||
logger: | ||
line: "152" | ||
name: "ngc_injector.py" | ||
message: "Valid profile: e45b4b991bbc51d0df3ce53e87060fc3a7f76555406ed534a8479c6faa706987 (tensorrt_llm-a10g-bf16-tp4-latency) on GPUs [0, 1, 2, 3]" | ||
status: "info" | ||
tags: | ||
- "source:LOGS_SOURCE" | ||
timestamp: 1730325388831 | ||
- | ||
sample: |- | ||
WARNING 2024-10-30 21:58:27.670 arg_utils.py:775] Chunked prefill is enabled by default for models with max_model_len > 32K. Currently, chunked prefill might not work with some features or models. If you encounter any issues, please disable chunked prefill by setting --enable-chunked-prefill=False. | ||
result: | ||
custom: | ||
level: "WARNING" | ||
timestamp: 1730325507670 | ||
logger: | ||
line: "775" | ||
name: "arg_utils.py" | ||
message: "Chunked prefill is enabled by default for models with max_model_len > 32K. Currently, chunked prefill might not work with some features or models. If you encounter any issues, please disable chunked prefill by setting --enable-chunked-prefill=False." | ||
status: "warn" | ||
tags: | ||
- "source:LOGS_SOURCE" | ||
timestamp: 1730325507670 |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,20 @@ | ||
{ | ||
"name": "NVIDIA NIM Errors", | ||
"options": { | ||
"columns": [ | ||
"host", | ||
"service" | ||
], | ||
"message_display": "inline", | ||
"show_date_column": true, | ||
"show_message_column": true, | ||
"show_timeline": true | ||
}, | ||
"page": "stream", | ||
"query": "source:nvidia_nim status:error", | ||
"timerange": { | ||
"interval_ms": 900000 | ||
}, | ||
"type": "logs", | ||
"visible_facets": [] | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
You can use the operator notSpace instead of data for this case (better performance)