Nh/unify request attribute extraction #45088
Annotations
9 errors and 6 warnings
|
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(91 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "008788d15a01152b",
+ "span_id": "282cb52a5cba2781",
+ "start_timestamp": 1774624710.488,
"status": "ok",
+ "timestamp": 1774624710.5886533,
+ "trace_id": "977fd99b7720ef74e81d40102e60bb62",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37595,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37595,
+ "url": "http://localhost:37595/openai/chat/completions",
+ "url.full": "http://localhost:37595/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:37595/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "282cb52a5cba2781",
+ "span_id": "308cbd103b3abce5",
+ "start_timestamp": 1774624710.524,
+ "status": "ok",
+ "timestamp": 1774624710.5797377,
+ "trace_id": "977fd99b7720ef74e81d40102e60bb62",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "008788d15a01152b",
+ "span_id": "9814b8ab999c3913",
+ "start_timestamp": 1774624710.589,
"status": "ok",
+ "timestamp": 1774624710.6125958,
+ "trace_id": "977fd99b7720ef74e81d40102e60bb62",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37595,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37595,
+ "url": "http://localhost:37595/openai/responses",
+ "url.full": "http://localhost:37595/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:37595/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "9814b8ab99
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(91 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "d28027a802ff7216",
+ "span_id": "c7b1a6594774752b",
+ "start_timestamp": 1774624706.873,
"status": "ok",
+ "timestamp": 1774624706.952055,
+ "trace_id": "a11666e263399d814754710a0a8d685d",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45413,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45413,
+ "url": "http://localhost:45413/openai/chat/completions",
+ "url.full": "http://localhost:45413/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:45413/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c7b1a6594774752b",
+ "span_id": "ffc37fa93ce18baa",
+ "start_timestamp": 1774624706.887,
+ "status": "ok",
+ "timestamp": 1774624706.948937,
+ "trace_id": "a11666e263399d814754710a0a8d685d",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "d28027a802ff7216",
+ "span_id": "c15f486329f011c6",
+ "start_timestamp": 1774624706.953,
"status": "ok",
+ "timestamp": 1774624706.9686575,
+ "trace_id": "a11666e263399d814754710a0a8d685d",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45413,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45413,
+ "url": "http://localhost:45413/openai/responses",
+ "url.full": "http://localhost:45413/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:45413/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c15f486329f0
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(91 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "8fa9a9c73f5a7226",
+ "span_id": "c618c564e1aa3d97",
+ "start_timestamp": 1774624700.931,
"status": "ok",
+ "timestamp": 1774624701.0713634,
+ "trace_id": "78907d90977229a1ddcf2464f5e5657e",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45875,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45875,
+ "url": "http://localhost:45875/openai/chat/completions",
+ "url.full": "http://localhost:45875/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:45875/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c618c564e1aa3d97",
+ "span_id": "e56943a73f317855",
+ "start_timestamp": 1774624700.974,
+ "status": "ok",
+ "timestamp": 1774624701.0673876,
+ "trace_id": "78907d90977229a1ddcf2464f5e5657e",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "8fa9a9c73f5a7226",
+ "span_id": "3aed547976e1fd99",
+ "start_timestamp": 1774624701.072,
"status": "ok",
+ "timestamp": 1774624701.091697,
+ "trace_id": "78907d90977229a1ddcf2464f5e5657e",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45875,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45875,
+ "url": "http://localhost:45875/openai/responses",
+ "url.full": "http://localhost:45875/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:45875/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "3aed547976e1fd99",
+ "span_id": "46bb967124224b16",
+ "start_timestamp": 1774624701.074,
+ "status": "ok",
+ "timestamp": 1774624701.089959,
+ "trace_id": "789
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(91 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "dddf31205b55ddf7",
+ "span_id": "2c7972ab08e1b130",
+ "start_timestamp": 1774624697.173,
"status": "ok",
+ "timestamp": 1774624697.2592187,
+ "trace_id": "925bd229d152383051a90825d95f5eb3",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 36939,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 36939,
+ "url": "http://localhost:36939/openai/chat/completions",
+ "url.full": "http://localhost:36939/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:36939/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "2c7972ab08e1b130",
+ "span_id": "8eb7abb68da8c340",
+ "start_timestamp": 1774624697.19,
+ "status": "ok",
+ "timestamp": 1774624697.255056,
+ "trace_id": "925bd229d152383051a90825d95f5eb3",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "dddf31205b55ddf7",
+ "span_id": "40e63675dc8af353",
+ "start_timestamp": 1774624697.26,
"status": "ok",
+ "timestamp": 1774624697.2756588,
+ "trace_id": "925bd229d152383051a90825d95f5eb3",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 36939,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 36939,
+ "url": "http://localhost:36939/openai/responses",
+ "url.full": "http://localhost:36939/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:36939/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "40e63675dc8af353",
+ "span_id": "51f45cefec9a0718",
+ "start_timestamp": 1774624697.262,
+ "status": "ok",
+ "timestamp": 1774624697.2742648,
+ "trace_id": "925b
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(91 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "70dba9dd7e1d2cb3",
+ "span_id": "f6dc9ed19c61e9db",
+ "start_timestamp": 1774624650.307,
"status": "ok",
+ "timestamp": 1774624650.4057324,
+ "trace_id": "5aba6ca8d86125f90bc7f712235a83e1",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 32781,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 32781,
+ "url": "http://localhost:32781/openai/chat/completions",
+ "url.full": "http://localhost:32781/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:32781/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "f6dc9ed19c61e9db",
+ "span_id": "9881af5ea1d3dd14",
+ "start_timestamp": 1774624650.342,
+ "status": "ok",
+ "timestamp": 1774624650.4025006,
+ "trace_id": "5aba6ca8d86125f90bc7f712235a83e1",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "70dba9dd7e1d2cb3",
+ "span_id": "ed53dce3ad258f3c",
+ "start_timestamp": 1774624650.406,
"status": "ok",
+ "timestamp": 1774624650.4181151,
+ "trace_id": "5aba6ca8d86125f90bc7f712235a83e1",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 32781,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 32781,
+ "url": "http://localhost:32781/openai/responses",
+ "url.full": "http://localhost:32781/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:32781/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "ed53dce3
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(91 matching properties omitted from actual)
- Expected
+ Received
@@ -1,12 +1,13 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -21,13 +22,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "5e1b81ff7901eb23",
+ "span_id": "38fe1321ecf3d950",
+ "start_timestamp": 1774624646.472,
"status": "ok",
+ "timestamp": 1774624646.629075,
+ "trace_id": "ee6f8792636c2fe87e71a5e61364141c",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40981,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40981,
+ "url": "http://localhost:40981/openai/chat/completions",
+ "url.full": "http://localhost:40981/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40981/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "38fe1321ecf3d950",
+ "span_id": "bd3140925ff06bad",
+ "start_timestamp": 1774624646.489,
+ "status": "ok",
+ "timestamp": 1774624646.6180537,
+ "trace_id": "ee6f8792636c2fe87e71a5e61364141c",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -43,13 +78,47 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "5e1b81ff7901eb23",
+ "span_id": "a39c157c57e7a886",
+ "start_timestamp": 1774624646.63,
"status": "ok",
+ "timestamp": 1774624646.654112,
+ "trace_id": "ee6f8792636c2fe87e71a5e61364141c",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40981,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40981,
+ "url": "http://localhost:40981/openai/responses",
+ "url.full": "http://localhost:40981/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40981/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "a39c157c57e
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(91 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "59d14a5267fa3bb1",
+ "span_id": "b46003eeb6d9e11f",
+ "start_timestamp": 1774624642.543,
"status": "ok",
+ "timestamp": 1774624642.6423454,
+ "trace_id": "d55f85d6d21e9ac1e45d39c617258042",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38893,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38893,
+ "url": "http://localhost:38893/openai/chat/completions",
+ "url.full": "http://localhost:38893/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38893/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b46003eeb6d9e11f",
+ "span_id": "31ddc9bf9b6da549",
+ "start_timestamp": 1774624642.569,
+ "status": "ok",
+ "timestamp": 1774624642.6300755,
+ "trace_id": "d55f85d6d21e9ac1e45d39c617258042",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "59d14a5267fa3bb1",
+ "span_id": "e3865e8d47836af8",
+ "start_timestamp": 1774624642.643,
"status": "ok",
+ "timestamp": 1774624642.6646125,
+ "trace_id": "d55f85d6d21e9ac1e45d39c617258042",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38893,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38893,
+ "url": "http://localhost:38893/openai/responses",
+ "url.full": "http://localhost:38893/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38893/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e3865e8d47836af8",
+ "span_id": "8b41715bbfbfacb3",
+ "start_timestamp": 1774624642.646,
+ "status": "ok",
+ "timestamp": 1774624642.6620722,
+ "trace_id":
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(91 matching properties omitted from actual)
- Expected
+ Received
@@ -1,11 +1,12 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
+ "gen_ai.request.max_tokens": 100,
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
"gen_ai.response.id": "chatcmpl-mock123",
"gen_ai.response.model": "gpt-3.5-turbo",
@@ -17,13 +18,47 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "7ecb0429a0902f21",
+ "span_id": "59d1574e9fb15183",
+ "start_timestamp": 1774624638.727,
"status": "ok",
+ "timestamp": 1774624638.7919652,
+ "trace_id": "16e4db91c4cea2e5a977c2f6711e0839",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 43689,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43689,
+ "url": "http://localhost:43689/openai/chat/completions",
+ "url.full": "http://localhost:43689/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:43689/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "59d1574e9fb15183",
+ "span_id": "9bdaf57016f76a2a",
+ "start_timestamp": 1774624638.739,
+ "status": "ok",
+ "timestamp": 1774624638.7886164,
+ "trace_id": "16e4db91c4cea2e5a977c2f6711e0839",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -36,26 +71,152 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "7ecb0429a0902f21",
+ "span_id": "0e58bdbfb3f0e4f8",
+ "start_timestamp": 1774624638.792,
"status": "ok",
+ "timestamp": 1774624638.8012319,
+ "trace_id": "16e4db91c4cea2e5a977c2f6711e0839",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 43689,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43689,
+ "url": "http://localhost:43689/openai/responses",
+ "url.full": "http://localhost:43689/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:43689/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "0e58bdbfb3f0e4f8",
+ "span_id": "a7105487e61daa5f",
+ "start_timestamp": 1774624638.794,
+ "status": "ok",
+ "timestamp": 1774624638.8008027,
+ "trace_id":
|
|
Node.js 20 actions are deprecated. The following actions are running on Node.js 20 and may not work as expected: actions/cache/restore@v4, actions/download-artifact@v4. Actions will be forced to run with Node.js 24 by default starting June 2nd, 2026. Node.js 20 will be removed from the runner on September 16th, 2026. Please check if updated versions of these actions are available that support Node.js 24. To opt into Node.js 24 now, set the FORCE_JAVASCRIPT_ACTIONS_TO_NODE24=true environment variable on the runner or in your workflow file. Once Node.js 24 becomes the default, you can temporarily opt out by setting ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION=true. For more information see: https://github.blog/changelog/2025-09-19-deprecation-of-node-20-on-github-actions-runners/
|
|
|
|
|
|
|
|
|
|
|
The logs for this run have expired and are no longer available.
Loading