hash
stringlengths
40
40
date
stringdate
2017-08-30 22:37:25
2025-03-22 03:08:47
author
stringclasses
173 values
commit_message
stringlengths
15
151
is_merge
bool
1 class
masked_commit_message
stringlengths
6
126
type
stringclasses
17 values
git_diff
stringlengths
182
1.51M
d48d3d6bf91b1f1385e4b0ea47ca070bfe83b001
2022-05-03 05:20:31
Vu Luong
ref(pageFIlter): Left-align trigger button text (#34150)
false
Left-align trigger button text (#34150)
ref
diff --git a/static/app/components/datePageFilter.tsx b/static/app/components/datePageFilter.tsx index f9c34e2f44f3ed..6f99421ebfb6cf 100644 --- a/static/app/components/datePageFilter.tsx +++ b/static/app/components/datePageFilter.tsx @@ -92,6 +92,7 @@ const TitleContainer = styled('div')` text-overflow: ellipsis; flex: 1 1 0%; margin-left: ${space(1)}; + text-align: left; `; const DropdownTitle = styled('div')` diff --git a/static/app/components/environmentPageFilter.tsx b/static/app/components/environmentPageFilter.tsx index c99bdc11d5a256..75676bc2a50c9c 100644 --- a/static/app/components/environmentPageFilter.tsx +++ b/static/app/components/environmentPageFilter.tsx @@ -111,7 +111,7 @@ function EnvironmentPageFilter({ const TitleContainer = styled('div')` display: flex; align-items: center; - justify-content: center; + justify-content: flex-start; flex: 1 1 0%; margin-left: ${space(1)}; overflow: hidden; diff --git a/static/app/components/projectPageFilter.tsx b/static/app/components/projectPageFilter.tsx index 71578d67a76bcd..785ea454535cd1 100644 --- a/static/app/components/projectPageFilter.tsx +++ b/static/app/components/projectPageFilter.tsx @@ -205,6 +205,7 @@ const TitleContainer = styled('div')` text-overflow: ellipsis; flex: 1 1 0%; margin-left: ${space(1)}; + text-align: left; `; const DropdownTitle = styled('div')`
3ba3b1553afae973abba17cbbff83d89d120deea
2023-04-21 23:35:32
Mark Story
fix: Fix actor_id constraint migration to account for duplicates (#47727)
false
Fix actor_id constraint migration to account for duplicates (#47727)
fix
diff --git a/src/sentry/migrations/0418_add_actor_constraints.py b/src/sentry/migrations/0418_add_actor_constraints.py index 82ccdfabfe794f..cc049df03e4802 100644 --- a/src/sentry/migrations/0418_add_actor_constraints.py +++ b/src/sentry/migrations/0418_add_actor_constraints.py @@ -24,6 +24,24 @@ class Migration(CheckedMigration): operations = ( [ + migrations.RunSQL( + sql=""" +WITH duplicate_users AS ( + SELECT user_id FROM sentry_actor WHERE user_id IS NOT NULL GROUP BY user_id HAVING count(*) > 1 +), +cleanup_actors AS ( + SELECT actor.id AS id FROM sentry_actor AS actor + LEFT JOIN auth_user AS u ON u.actor_id = actor.id + WHERE actor.user_id IN (SELECT user_id FROM duplicate_users) + AND u.actor_id IS NULL +) +UPDATE sentry_actor SET user_id = NULL WHERE id IN (SELECT id FROM cleanup_actors); + """, + reverse_sql="", + hints={"tables": ["sentry_actor"]}, + ), + ] + + [ migrations.RunSQL( sql=line, reverse_sql="",
4868f8e377c90103a3547d9ee7663fdfd266a446
2024-08-16 23:51:05
Christinarlong
ref(control_silo): Move Identitiy model to users module (#76272)
false
Move Identitiy model to users module (#76272)
ref
diff --git a/fixtures/gitlab.py b/fixtures/gitlab.py index a2b800f6068642..0ce0948e070d7b 100644 --- a/fixtures/gitlab.py +++ b/fixtures/gitlab.py @@ -2,12 +2,12 @@ from sentry.integrations.gitlab.integration import GitlabIntegration from sentry.integrations.models.integration import Integration -from sentry.models.identity import Identity, IdentityProvider from sentry.models.repository import Repository from sentry.silo.base import SiloMode from sentry.testutils.cases import APITestCase from sentry.testutils.helpers.integrations import get_installation_of_type from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity, IdentityProvider EXTERNAL_ID = "example.gitlab.com:group-x" WEBHOOK_SECRET = "secret-token-value" diff --git a/src/sentry/api/endpoints/user_identity.py b/src/sentry/api/endpoints/user_identity.py index d41803832aa8ab..b45e8a557a26ef 100644 --- a/src/sentry/api/endpoints/user_identity.py +++ b/src/sentry/api/endpoints/user_identity.py @@ -6,7 +6,7 @@ from sentry.api.bases.user import UserEndpoint from sentry.api.paginator import OffsetPaginator from sentry.api.serializers import serialize -from sentry.models.identity import Identity +from sentry.users.models.identity import Identity @control_silo_endpoint diff --git a/src/sentry/api/endpoints/user_identity_config.py b/src/sentry/api/endpoints/user_identity_config.py index f43f2b3a5ccb8d..c7bcb432eedfa8 100644 --- a/src/sentry/api/endpoints/user_identity_config.py +++ b/src/sentry/api/endpoints/user_identity_config.py @@ -16,7 +16,7 @@ supports_login, ) from sentry.models.authidentity import AuthIdentity -from sentry.models.identity import Identity +from sentry.users.models.identity import Identity from sentry.users.models.user import User from social_auth.models import UserSocialAuth diff --git a/src/sentry/api/serializers/models/identity.py b/src/sentry/api/serializers/models/identity.py index 792ab9aedd0554..189c2baf620c1d 100644 --- a/src/sentry/api/serializers/models/identity.py +++ b/src/sentry/api/serializers/models/identity.py @@ -1,5 +1,5 @@ from sentry.api.serializers import Serializer, register, serialize -from sentry.models.identity import Identity +from sentry.users.models.identity import Identity @register(Identity) diff --git a/src/sentry/api/serializers/models/identityprovider.py b/src/sentry/api/serializers/models/identityprovider.py index d9a01445962ac9..70c61ea47f0701 100644 --- a/src/sentry/api/serializers/models/identityprovider.py +++ b/src/sentry/api/serializers/models/identityprovider.py @@ -1,5 +1,5 @@ from sentry.api.serializers import Serializer, register -from sentry.models.identity import IdentityProvider +from sentry.users.models.identity import IdentityProvider @register(IdentityProvider) diff --git a/src/sentry/api/serializers/models/user_identity_config.py b/src/sentry/api/serializers/models/user_identity_config.py index 4f5f7132f6a975..749a67db7e0ce8 100644 --- a/src/sentry/api/serializers/models/user_identity_config.py +++ b/src/sentry/api/serializers/models/user_identity_config.py @@ -16,7 +16,7 @@ from sentry.identity import is_login_provider from sentry.integrations.manager import default_manager as integrations from sentry.models.authidentity import AuthIdentity -from sentry.models.identity import Identity +from sentry.users.models.identity import Identity from social_auth.models import UserSocialAuth from . import user_social_auth diff --git a/src/sentry/deletions/defaults/organizationintegration.py b/src/sentry/deletions/defaults/organizationintegration.py index ef86b8746ae2ca..941dcee637857e 100644 --- a/src/sentry/deletions/defaults/organizationintegration.py +++ b/src/sentry/deletions/defaults/organizationintegration.py @@ -11,7 +11,7 @@ def should_proceed(self, instance): return instance.status in {ObjectStatus.DELETION_IN_PROGRESS, ObjectStatus.PENDING_DELETION} def get_child_relations(self, instance): - from sentry.models.identity import Identity + from sentry.users.models.identity import Identity relations = [] diff --git a/src/sentry/identity/pipeline.py b/src/sentry/identity/pipeline.py index a5efea9f8e1957..1e4f8f3b21f5cb 100644 --- a/src/sentry/identity/pipeline.py +++ b/src/sentry/identity/pipeline.py @@ -5,8 +5,8 @@ from django.urls import reverse from django.utils.translation import gettext_lazy as _ -from sentry.models.identity import Identity, IdentityProvider from sentry.pipeline import Pipeline +from sentry.users.models.identity import Identity, IdentityProvider from sentry.utils import metrics from . import default_manager diff --git a/src/sentry/identity/services/identity/impl.py b/src/sentry/identity/services/identity/impl.py index e1687dda791555..73e4110ea259fd 100644 --- a/src/sentry/identity/services/identity/impl.py +++ b/src/sentry/identity/services/identity/impl.py @@ -15,7 +15,7 @@ from sentry.identity.services.identity.serial import serialize_identity, serialize_identity_provider from sentry.identity.services.identity.service import IdentityService from sentry.models.authidentity import AuthIdentity -from sentry.models.identity import Identity +from sentry.users.models.identity import Identity class DatabaseBackedIdentityService(IdentityService): @@ -26,7 +26,7 @@ def get_provider( provider_type: str | None = None, provider_ext_id: str | None = None, ) -> RpcIdentityProvider | None: - from sentry.models.identity import IdentityProvider + from sentry.users.models.identity import IdentityProvider # If an id is provided, use that -- otherwise, use the type and external_id idp_kwargs: Any = ( @@ -57,7 +57,7 @@ def get_user_identities_by_provider_type( ) -> list[RpcIdentity]: from django.db.models import F - from sentry.models.identity import Identity + from sentry.users.models.identity import Identity identities = Identity.objects.filter(user=user_id, idp__type=provider_type) diff --git a/src/sentry/identity/services/identity/model.py b/src/sentry/identity/services/identity/model.py index ac3277c8aca28f..24c0115ab104b6 100644 --- a/src/sentry/identity/services/identity/model.py +++ b/src/sentry/identity/services/identity/model.py @@ -28,7 +28,7 @@ class RpcIdentity(RpcModel): def get_identity(self) -> "Provider": from sentry.identity import get from sentry.identity.services.identity import identity_service - from sentry.models.identity import IdentityProvider + from sentry.users.models.identity import IdentityProvider identity_provider = identity_service.get_provider(provider_id=self.idp_id) if identity_provider is None: diff --git a/src/sentry/identity/services/identity/serial.py b/src/sentry/identity/services/identity/serial.py index 61dec5feebe219..e8afbd38c1914e 100644 --- a/src/sentry/identity/services/identity/serial.py +++ b/src/sentry/identity/services/identity/serial.py @@ -3,7 +3,7 @@ from sentry.identity.services.identity import RpcIdentity, RpcIdentityProvider if TYPE_CHECKING: - from sentry.models.identity import Identity, IdentityProvider + from sentry.users.models.identity import Identity, IdentityProvider def serialize_identity_provider(identity_provider: "IdentityProvider") -> RpcIdentityProvider: diff --git a/src/sentry/integrations/base.py b/src/sentry/integrations/base.py index 2dab3fb12a8844..abd91547629fae 100644 --- a/src/sentry/integrations/base.py +++ b/src/sentry/integrations/base.py @@ -20,7 +20,6 @@ from sentry.integrations.models.integration import Integration from sentry.integrations.notify_disable import notify_disable from sentry.integrations.request_buffer import IntegrationRequestBuffer -from sentry.models.identity import Identity from sentry.models.team import Team from sentry.organizations.services.organization import ( RpcOrganization, @@ -42,6 +41,7 @@ IntegrationFormError, UnsupportedResponseType, ) +from sentry.users.models.identity import Identity from sentry.utils.audit import create_audit_entry, create_system_audit_entry from sentry.utils.sdk import Scope diff --git a/src/sentry/integrations/bitbucket_server/integration.py b/src/sentry/integrations/bitbucket_server/integration.py index 3fadf550245c59..01ab5b7ecadd63 100644 --- a/src/sentry/integrations/bitbucket_server/integration.py +++ b/src/sentry/integrations/bitbucket_server/integration.py @@ -26,11 +26,11 @@ from sentry.integrations.services.repository.model import RpcRepository from sentry.integrations.source_code_management.repository import RepositoryIntegration from sentry.integrations.tasks.migrate_repo import migrate_repo -from sentry.models.identity import Identity from sentry.models.repository import Repository from sentry.organizations.services.organization import RpcOrganizationSummary from sentry.pipeline import PipelineView from sentry.shared_integrations.exceptions import ApiError, IntegrationError +from sentry.users.models.identity import Identity from sentry.web.helpers import render_to_response from .client import BitbucketServerClient, BitbucketServerSetupClient diff --git a/src/sentry/integrations/discord/views/link_identity.py b/src/sentry/integrations/discord/views/link_identity.py index a1f139f18dfa00..3bb1bdf4a2b188 100644 --- a/src/sentry/integrations/discord/views/link_identity.py +++ b/src/sentry/integrations/discord/views/link_identity.py @@ -8,8 +8,8 @@ from sentry.integrations.services.integration.model import RpcIntegration from sentry.integrations.types import ExternalProviders from sentry.integrations.utils.identities import get_identity_or_404 -from sentry.models.identity import Identity from sentry.types.actor import ActorType +from sentry.users.models.identity import Identity from sentry.utils.http import absolute_uri from sentry.utils.signing import sign, unsign from sentry.web.frontend.base import BaseView, control_silo_view diff --git a/src/sentry/integrations/discord/views/unlink_identity.py b/src/sentry/integrations/discord/views/unlink_identity.py index 019efa20633d77..00dc3c375ccfad 100644 --- a/src/sentry/integrations/discord/views/unlink_identity.py +++ b/src/sentry/integrations/discord/views/unlink_identity.py @@ -9,8 +9,8 @@ from sentry.integrations.services.integration.model import RpcIntegration from sentry.integrations.types import ExternalProviders from sentry.integrations.utils.identities import get_identity_or_404 -from sentry.models.identity import Identity from sentry.types.actor import ActorType +from sentry.users.models.identity import Identity from sentry.utils.http import absolute_uri from sentry.utils.signing import sign, unsign from sentry.web.frontend.base import BaseView, control_silo_view diff --git a/src/sentry/integrations/gitlab/integration.py b/src/sentry/integrations/gitlab/integration.py index 210bfdcab5fcb7..fbb1efee2332f3 100644 --- a/src/sentry/integrations/gitlab/integration.py +++ b/src/sentry/integrations/gitlab/integration.py @@ -19,7 +19,6 @@ from sentry.integrations.services.repository.model import RpcRepository from sentry.integrations.source_code_management.commit_context import CommitContextIntegration from sentry.integrations.source_code_management.repository import RepositoryIntegration -from sentry.models.identity import Identity from sentry.models.repository import Repository from sentry.pipeline import NestedPipelineView, PipelineView from sentry.shared_integrations.exceptions import ( @@ -27,6 +26,7 @@ IntegrationError, IntegrationProviderError, ) +from sentry.users.models.identity import Identity from sentry.utils.hashlib import sha1_text from sentry.utils.http import absolute_uri from sentry.web.helpers import render_to_response diff --git a/src/sentry/integrations/jira_server/integration.py b/src/sentry/integrations/jira_server/integration.py index fe0059f86f5b65..6472183c0dbdd5 100644 --- a/src/sentry/integrations/jira_server/integration.py +++ b/src/sentry/integrations/jira_server/integration.py @@ -32,7 +32,6 @@ from sentry.integrations.models.integration_external_project import IntegrationExternalProject from sentry.integrations.services.integration import integration_service from sentry.models.group import Group -from sentry.models.identity import Identity from sentry.organizations.services.organization.service import organization_service from sentry.pipeline import PipelineView from sentry.shared_integrations.exceptions import ( @@ -43,6 +42,7 @@ IntegrationFormError, ) from sentry.silo.base import all_silo_function +from sentry.users.models.identity import Identity from sentry.users.services.user import RpcUser from sentry.users.services.user.service import user_service from sentry.utils.hashlib import sha1_text diff --git a/src/sentry/integrations/msteams/link_identity.py b/src/sentry/integrations/msteams/link_identity.py index 1d82c994916e50..b5fbfce02fee4b 100644 --- a/src/sentry/integrations/msteams/link_identity.py +++ b/src/sentry/integrations/msteams/link_identity.py @@ -7,7 +7,7 @@ from sentry.integrations.types import ExternalProviders from sentry.integrations.utils import get_identity_or_404 -from sentry.models.identity import Identity +from sentry.users.models.identity import Identity from sentry.utils.http import absolute_uri from sentry.utils.signing import sign, unsign from sentry.web.frontend.base import BaseView, control_silo_view diff --git a/src/sentry/integrations/msteams/unlink_identity.py b/src/sentry/integrations/msteams/unlink_identity.py index 7a8465aab73242..b70ced69f9a9a8 100644 --- a/src/sentry/integrations/msteams/unlink_identity.py +++ b/src/sentry/integrations/msteams/unlink_identity.py @@ -4,7 +4,7 @@ from django.utils.decorators import method_decorator from django.views.decorators.cache import never_cache -from sentry.models.identity import Identity +from sentry.users.models.identity import Identity from sentry.utils.http import absolute_uri from sentry.utils.signing import sign, unsign from sentry.web.frontend.base import BaseView, control_silo_view diff --git a/src/sentry/integrations/pipeline.py b/src/sentry/integrations/pipeline.py index 12d9aad62690b5..34975d7c914b9c 100644 --- a/src/sentry/integrations/pipeline.py +++ b/src/sentry/integrations/pipeline.py @@ -13,12 +13,12 @@ from sentry.integrations.manager import default_manager from sentry.integrations.models.integration import Integration from sentry.integrations.models.organization_integration import OrganizationIntegration -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.models.organizationmapping import OrganizationMapping from sentry.organizations.absolute_url import generate_organization_url from sentry.pipeline import Pipeline, PipelineAnalyticsEntry from sentry.shared_integrations.exceptions import IntegrationError, IntegrationProviderError from sentry.silo.base import SiloMode +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.web.helpers import render_to_response __all__ = ["IntegrationPipeline"] diff --git a/src/sentry/integrations/slack/tasks/link_slack_user_identities.py b/src/sentry/integrations/slack/tasks/link_slack_user_identities.py index 805ee9b9c01219..14173835343305 100644 --- a/src/sentry/integrations/slack/tasks/link_slack_user_identities.py +++ b/src/sentry/integrations/slack/tasks/link_slack_user_identities.py @@ -8,10 +8,10 @@ from sentry.integrations.services.integration import integration_service from sentry.integrations.slack.utils.users import SlackUserData, get_slack_data_by_user from sentry.integrations.utils import get_identities_by_user -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.organizations.services.organization import organization_service from sentry.silo.base import SiloMode from sentry.tasks.base import instrumented_task +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.users.models.user import User from sentry.users.models.useremail import UserEmail diff --git a/src/sentry/integrations/slack/views/link_identity.py b/src/sentry/integrations/slack/views/link_identity.py index 09143a99dcc7eb..8757c49e2f7bfd 100644 --- a/src/sentry/integrations/slack/views/link_identity.py +++ b/src/sentry/integrations/slack/views/link_identity.py @@ -18,9 +18,9 @@ from sentry.integrations.slack.views.types import IdentityParams from sentry.integrations.types import ExternalProviderEnum, ExternalProviders from sentry.integrations.utils import get_identity_or_404 -from sentry.models.identity import Identity from sentry.notifications.notificationcontroller import NotificationController from sentry.notifications.notifications.integration_nudge import IntegrationNudgeNotification +from sentry.users.models.identity import Identity from sentry.utils import metrics from sentry.utils.signing import unsign from sentry.web.frontend.base import BaseView, control_silo_view diff --git a/src/sentry/integrations/slack/views/types.py b/src/sentry/integrations/slack/views/types.py index 1d43b7193f8c3e..24b32299062393 100644 --- a/src/sentry/integrations/slack/views/types.py +++ b/src/sentry/integrations/slack/views/types.py @@ -1,8 +1,8 @@ from dataclasses import dataclass from sentry.integrations.models.integration import Integration -from sentry.models.identity import IdentityProvider from sentry.organizations.services.organization import RpcOrganization +from sentry.users.models.identity import IdentityProvider @dataclass(frozen=True) diff --git a/src/sentry/integrations/slack/views/unlink_identity.py b/src/sentry/integrations/slack/views/unlink_identity.py index 275c1ed0bbabff..ce93e9ab90ef29 100644 --- a/src/sentry/integrations/slack/views/unlink_identity.py +++ b/src/sentry/integrations/slack/views/unlink_identity.py @@ -17,7 +17,7 @@ from sentry.integrations.slack.views.types import IdentityParams from sentry.integrations.types import ExternalProviders from sentry.integrations.utils import get_identity_or_404 -from sentry.models.identity import Identity +from sentry.users.models.identity import Identity from sentry.utils import metrics from sentry.utils.signing import unsign from sentry.web.frontend.base import BaseView, control_silo_view diff --git a/src/sentry/integrations/source_code_management/commit_context.py b/src/sentry/integrations/source_code_management/commit_context.py index a36a8d1e439bbe..61086249f1b359 100644 --- a/src/sentry/integrations/source_code_management/commit_context.py +++ b/src/sentry/integrations/source_code_management/commit_context.py @@ -8,8 +8,8 @@ from sentry.auth.exceptions import IdentityNotValid from sentry.integrations.models.repository_project_path_config import RepositoryProjectPathConfig -from sentry.models.identity import Identity from sentry.models.repository import Repository +from sentry.users.models.identity import Identity @dataclass diff --git a/src/sentry/integrations/source_code_management/repository.py b/src/sentry/integrations/source_code_management/repository.py index 58ff93257d4f5f..2efb81af154c99 100644 --- a/src/sentry/integrations/source_code_management/repository.py +++ b/src/sentry/integrations/source_code_management/repository.py @@ -9,10 +9,10 @@ from sentry.auth.exceptions import IdentityNotValid from sentry.integrations.base import IntegrationInstallation from sentry.integrations.services.repository import RpcRepository -from sentry.models.identity import Identity from sentry.models.repository import Repository from sentry.shared_integrations.client.base import BaseApiResponseX from sentry.shared_integrations.exceptions import ApiError, IntegrationError +from sentry.users.models.identity import Identity from sentry.utils import metrics REPOSITORY_INTEGRATION_CHECK_FILE_METRIC = "repository_integration.check_file.{result}" diff --git a/src/sentry/integrations/utils/identities.py b/src/sentry/integrations/utils/identities.py index d1917751a7fb5d..7d93614d5ddc62 100644 --- a/src/sentry/integrations/utils/identities.py +++ b/src/sentry/integrations/utils/identities.py @@ -7,9 +7,9 @@ from sentry.integrations.models.integration import Integration from sentry.integrations.models.organization_integration import OrganizationIntegration from sentry.integrations.types import EXTERNAL_PROVIDERS, ExternalProviders -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.organizations.services.organization import RpcOrganization, organization_service from sentry.silo.base import control_silo_function +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.users.models.user import User from sentry.users.services.user.service import user_service diff --git a/src/sentry/integrations/vsts/client.py b/src/sentry/integrations/vsts/client.py index c8de055a9c3599..6d419aaa0e15e0 100644 --- a/src/sentry/integrations/vsts/client.py +++ b/src/sentry/integrations/vsts/client.py @@ -12,11 +12,11 @@ from sentry.integrations.base import IntegrationFeatureNotImplementedError from sentry.integrations.client import ApiClient from sentry.integrations.source_code_management.repository import RepositoryClient -from sentry.models.identity import Identity from sentry.models.repository import Repository from sentry.shared_integrations.client.base import BaseApiResponseX from sentry.shared_integrations.client.proxy import IntegrationProxyClient from sentry.silo.base import control_silo_function +from sentry.users.models.identity import Identity from sentry.utils.http import absolute_uri if TYPE_CHECKING: diff --git a/src/sentry/models/__init__.py b/src/sentry/models/__init__.py index 2018589287e6fd..2c39a564ffb87a 100644 --- a/src/sentry/models/__init__.py +++ b/src/sentry/models/__init__.py @@ -1,5 +1,6 @@ from sentry.users.models.authenticator import * # NOQA from sentry.users.models.email import * # NOQA +from sentry.users.models.identity import * # NOQA from sentry.users.models.lostpasswordhash import * # NOQA from sentry.users.models.user import * # NOQA from sentry.users.models.useremail import * # NOQA @@ -66,7 +67,6 @@ from .groupsnooze import * # NOQA from .groupsubscription import * # NOQA from .grouptombstone import * # NOQA -from .identity import * # NOQA from .importchunk import * # NOQA from .integrations import * # NOQA from .latestreporeleaseenvironment import * # NOQA diff --git a/src/sentry/models/identity.py b/src/sentry/models/identity.py index 26a335dbd9436a..68ab6f1c03128f 100644 --- a/src/sentry/models/identity.py +++ b/src/sentry/models/identity.py @@ -1,216 +1,3 @@ -from __future__ import annotations +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus -import logging -from collections.abc import Mapping -from typing import TYPE_CHECKING, Any, ClassVar - -from django.conf import settings -from django.db import IntegrityError, models -from django.db.models import Q, QuerySet -from django.utils import timezone - -from sentry import analytics -from sentry.backup.scopes import RelocationScope -from sentry.db.models import ( - ArrayField, - BoundedPositiveIntegerField, - FlexibleForeignKey, - Model, - control_silo_model, -) -from sentry.db.models.fields.jsonfield import JSONField -from sentry.db.models.manager.base import BaseManager -from sentry.integrations.types import ExternalProviders -from sentry.users.services.user import RpcUser - -if TYPE_CHECKING: - from sentry.identity.base import Provider - from sentry.identity.services.identity import RpcIdentityProvider - from sentry.users.models.user import User - -logger = logging.getLogger(__name__) - - -# TODO(dcramer): pull in enum library -class IdentityStatus: - UNKNOWN = 0 - VALID = 1 - INVALID = 2 - - -@control_silo_model -class IdentityProvider(Model): - """ - An IdentityProvider is an instance of a provider. - - The IdentityProvider is unique on the type of provider (eg github, slack, - google, etc). - - A SAML identity provide might look like this, type: onelogin, instance: - acme-org.onelogin.com. - """ - - __relocation_scope__ = RelocationScope.Excluded - - type = models.CharField(max_length=64) - config: models.Field[dict[str, Any], dict[str, Any]] = JSONField() - date_added = models.DateTimeField(default=timezone.now, null=True) - external_id = models.CharField(max_length=64, null=True) - - class Meta: - app_label = "sentry" - db_table = "sentry_identityprovider" - unique_together = (("type", "external_id"),) - - def get_provider(self): - from sentry.identity import get - - return get(self.type) - - -class IdentityManager(BaseManager["Identity"]): - def get_identities_for_user( - self, user: User | RpcUser, provider: ExternalProviders - ) -> QuerySet: - return self.filter(user_id=user.id, idp__type=provider.name) - - def has_identity(self, user: User | RpcUser, provider: ExternalProviders) -> bool: - return self.get_identities_for_user(user, provider).exists() - - def link_identity( - self, - user: User | RpcUser, - idp: IdentityProvider | RpcIdentityProvider, - external_id: str, - should_reattach: bool = True, - defaults: Mapping[str, Any | None] | None = None, - ) -> Identity: - """ - Link the user with the identity. If `should_reattach` is passed, handle - the case where the user is linked to a different identity or the - identity is linked to a different user. - """ - defaults = { - **(defaults or {}), - "status": IdentityStatus.VALID, - "date_verified": timezone.now(), - } - try: - identity, created = self.get_or_create( - idp_id=idp.id, user_id=user.id, external_id=external_id, defaults=defaults - ) - if not created: - identity.update(**defaults) - except IntegrityError: - if not should_reattach: - raise - return self.reattach(idp, external_id, user, defaults) - - analytics.record( - "integrations.identity_linked", - provider="slack", - # Note that prior to circa March 2023 this was user.actor_id. It changed - # when actor ids were no longer stable between regions for the same user - actor_id=user.id, - actor_type="user", - ) - return identity - - def delete_identity( - self, user: User | RpcUser, idp: IdentityProvider | RpcIdentityProvider, external_id: str - ) -> None: - self.filter(Q(external_id=external_id) | Q(user_id=user.id), idp_id=idp.id).delete() - logger.info( - "deleted-identity", - extra={"external_id": external_id, "idp_id": idp.id, "user_id": user.id}, - ) - - def create_identity( - self, - idp: IdentityProvider | RpcIdentityProvider, - external_id: str, - user: User | RpcUser, - defaults: Mapping[str, Any], - ) -> Identity: - identity_model = self.create( - idp_id=idp.id, user_id=user.id, external_id=external_id, **defaults - ) - logger.info( - "created-identity", - extra={ - "idp_id": idp.id, - "external_id": external_id, - "object_id": identity_model.id, - "user_id": user.id, - }, - ) - return identity_model - - def reattach( - self, - idp: IdentityProvider | RpcIdentityProvider, - external_id: str, - user: User | RpcUser, - defaults: Mapping[str, Any], - ) -> Identity: - """ - Removes identities under `idp` associated with either `external_id` or `user` - and creates a new identity linking them. - """ - self.delete_identity(user=user, idp=idp, external_id=external_id) - return self.create_identity(user=user, idp=idp, external_id=external_id, defaults=defaults) - - def update_external_id_and_defaults( - self, - idp: IdentityProvider, - external_id: str, - user: User | RpcUser, - defaults: Mapping[str, Any], - ) -> Identity: - """ - Updates the identity object for a given user and identity provider - with the new external id and other fields related to the identity status - """ - query = self.filter(user_id=user.id, idp=idp) - query.update(external_id=external_id, **defaults) - identity_model = query.get() - logger.info( - "updated-identity", - extra={ - "external_id": external_id, - "idp_id": idp.id, - "user_id": user.id, - "identity_id": identity_model.id, - }, - ) - return identity_model - - -@control_silo_model -class Identity(Model): - """ - A verified link between a user and a third party identity. - """ - - __relocation_scope__ = RelocationScope.Excluded - - idp = FlexibleForeignKey("sentry.IdentityProvider") - user = FlexibleForeignKey(settings.AUTH_USER_MODEL) - external_id = models.TextField() - data: models.Field[dict[str, Any], dict[str, Any]] = JSONField() - status = BoundedPositiveIntegerField(default=IdentityStatus.UNKNOWN) - scopes = ArrayField() - date_verified = models.DateTimeField(default=timezone.now) - date_added = models.DateTimeField(default=timezone.now) - - objects: ClassVar[IdentityManager] = IdentityManager() - - class Meta: - app_label = "sentry" - db_table = "sentry_identity" - unique_together = (("idp", "external_id"), ("idp", "user")) - - def get_provider(self) -> Provider: - from sentry.identity import get - - return get(self.idp.type) +__all__ = ("Identity", "IdentityProvider", "IdentityStatus") diff --git a/src/sentry/testutils/cases.py b/src/sentry/testutils/cases.py index 50c4d0c3a5b664..1ec11c1eb66d30 100644 --- a/src/sentry/testutils/cases.py +++ b/src/sentry/testutils/cases.py @@ -99,7 +99,6 @@ from sentry.models.environment import Environment from sentry.models.files.file import File from sentry.models.groupmeta import GroupMeta -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.models.notificationsettingoption import NotificationSettingOption from sentry.models.notificationsettingprovider import NotificationSettingProvider from sentry.models.options.project_option import ProjectOption @@ -143,6 +142,7 @@ from sentry.testutils.helpers.slack import install_slack from sentry.testutils.pytest.selenium import Browser from sentry.types.condition_activity import ConditionActivity, ConditionActivityType +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.users.models.user import User from sentry.users.models.user_option import UserOption from sentry.users.models.useremail import UserEmail diff --git a/src/sentry/testutils/factories.py b/src/sentry/testutils/factories.py index 4e5cff702ea91d..1d6339661d4a4b 100644 --- a/src/sentry/testutils/factories.py +++ b/src/sentry/testutils/factories.py @@ -101,7 +101,6 @@ from sentry.models.grouphistory import GroupHistory from sentry.models.grouplink import GroupLink from sentry.models.grouprelease import GroupRelease -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.models.integrations.sentry_app import SentryApp from sentry.models.integrations.sentry_app_installation import SentryAppInstallation from sentry.models.integrations.sentry_app_installation_for_provider import ( @@ -162,6 +161,7 @@ UptimeStatus, UptimeSubscription, ) +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.users.models.user import User from sentry.users.models.user_avatar import UserAvatar from sentry.users.models.user_option import UserOption diff --git a/src/sentry/testutils/fixtures.py b/src/sentry/testutils/fixtures.py index bfdc638508ab6f..cf7665c16e73ca 100644 --- a/src/sentry/testutils/fixtures.py +++ b/src/sentry/testutils/fixtures.py @@ -15,7 +15,6 @@ from sentry.integrations.models.organization_integration import OrganizationIntegration from sentry.models.activity import Activity from sentry.models.grouprelease import GroupRelease -from sentry.models.identity import Identity, IdentityProvider from sentry.models.organization import Organization from sentry.models.organizationmember import OrganizationMember from sentry.models.organizationmemberteam import OrganizationMemberTeam @@ -41,6 +40,7 @@ UptimeStatus, UptimeSubscription, ) +from sentry.users.models.identity import Identity, IdentityProvider from sentry.users.models.user import User from sentry.users.services.user import RpcUser diff --git a/src/sentry/testutils/helpers/slack.py b/src/sentry/testutils/helpers/slack.py index d19ea51ab832f4..000a0155d50827 100644 --- a/src/sentry/testutils/helpers/slack.py +++ b/src/sentry/testutils/helpers/slack.py @@ -7,11 +7,11 @@ from sentry.integrations.models.organization_integration import OrganizationIntegration from sentry.integrations.slack.message_builder.types import SlackBody from sentry.integrations.types import EXTERNAL_PROVIDERS, ExternalProviders -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.models.organization import Organization from sentry.models.team import Team from sentry.silo.base import SiloMode from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.users.models.user import User from sentry.utils import json diff --git a/src/sentry/users/models/identity.py b/src/sentry/users/models/identity.py new file mode 100644 index 00000000000000..cd9b9aba604cdd --- /dev/null +++ b/src/sentry/users/models/identity.py @@ -0,0 +1,216 @@ +from __future__ import annotations + +import logging +from collections.abc import Mapping +from typing import TYPE_CHECKING, Any, ClassVar + +from django.conf import settings +from django.db import IntegrityError, models +from django.db.models import Q, QuerySet +from django.utils import timezone + +from sentry import analytics +from sentry.backup.scopes import RelocationScope +from sentry.db.models import ( + ArrayField, + BoundedPositiveIntegerField, + FlexibleForeignKey, + Model, + control_silo_model, +) +from sentry.db.models.fields.jsonfield import JSONField +from sentry.db.models.manager.base import BaseManager +from sentry.integrations.types import ExternalProviders +from sentry.users.services.user import RpcUser + +if TYPE_CHECKING: + from sentry.identity.base import Provider + from sentry.identity.services.identity import RpcIdentityProvider + from sentry.users.models.user import User + +logger = logging.getLogger(__name__) + + +# TODO(dcramer): pull in enum library +class IdentityStatus: + UNKNOWN = 0 + VALID = 1 + INVALID = 2 + + +@control_silo_model +class IdentityProvider(Model): + """ + An IdentityProvider is an instance of a provider. + + The IdentityProvider is unique on the type of provider (eg github, slack, + google, etc). + + A SAML identity provide might look like this, type: onelogin, instance: + acme-org.onelogin.com. + """ + + __relocation_scope__ = RelocationScope.Excluded + + type = models.CharField(max_length=64) + config: models.Field[dict[str, Any], dict[str, Any]] = JSONField() + date_added = models.DateTimeField(default=timezone.now, null=True) + external_id = models.CharField(max_length=64, null=True) + + class Meta: + app_label = "sentry" + db_table = "sentry_identityprovider" + unique_together = (("type", "external_id"),) + + def get_provider(self) -> IdentityProvider: + from sentry.identity import get + + return get(self.type) + + +class IdentityManager(BaseManager["Identity"]): + def get_identities_for_user( + self, user: User | RpcUser, provider: ExternalProviders + ) -> QuerySet[Identity]: + return self.filter(user_id=user.id, idp__type=provider.name) + + def has_identity(self, user: User | RpcUser, provider: ExternalProviders) -> bool: + return self.get_identities_for_user(user, provider).exists() + + def link_identity( + self, + user: User | RpcUser, + idp: IdentityProvider | RpcIdentityProvider, + external_id: str, + should_reattach: bool = True, + defaults: Mapping[str, Any | None] | None = None, + ) -> Identity: + """ + Link the user with the identity. If `should_reattach` is passed, handle + the case where the user is linked to a different identity or the + identity is linked to a different user. + """ + defaults = { + **(defaults or {}), + "status": IdentityStatus.VALID, + "date_verified": timezone.now(), + } + try: + identity, created = self.get_or_create( + idp_id=idp.id, user_id=user.id, external_id=external_id, defaults=defaults + ) + if not created: + identity.update(**defaults) + except IntegrityError: + if not should_reattach: + raise + return self.reattach(idp, external_id, user, defaults) + + analytics.record( + "integrations.identity_linked", + provider="slack", + # Note that prior to circa March 2023 this was user.actor_id. It changed + # when actor ids were no longer stable between regions for the same user + actor_id=user.id, + actor_type="user", + ) + return identity + + def delete_identity( + self, user: User | RpcUser, idp: IdentityProvider | RpcIdentityProvider, external_id: str + ) -> None: + self.filter(Q(external_id=external_id) | Q(user_id=user.id), idp_id=idp.id).delete() + logger.info( + "deleted-identity", + extra={"external_id": external_id, "idp_id": idp.id, "user_id": user.id}, + ) + + def create_identity( + self, + idp: IdentityProvider | RpcIdentityProvider, + external_id: str, + user: User | RpcUser, + defaults: Mapping[str, Any], + ) -> Identity: + identity_model = self.create( + idp_id=idp.id, user_id=user.id, external_id=external_id, **defaults + ) + logger.info( + "created-identity", + extra={ + "idp_id": idp.id, + "external_id": external_id, + "object_id": identity_model.id, + "user_id": user.id, + }, + ) + return identity_model + + def reattach( + self, + idp: IdentityProvider | RpcIdentityProvider, + external_id: str, + user: User | RpcUser, + defaults: Mapping[str, Any], + ) -> Identity: + """ + Removes identities under `idp` associated with either `external_id` or `user` + and creates a new identity linking them. + """ + self.delete_identity(user=user, idp=idp, external_id=external_id) + return self.create_identity(user=user, idp=idp, external_id=external_id, defaults=defaults) + + def update_external_id_and_defaults( + self, + idp: IdentityProvider, + external_id: str, + user: User | RpcUser, + defaults: Mapping[str, Any], + ) -> Identity: + """ + Updates the identity object for a given user and identity provider + with the new external id and other fields related to the identity status + """ + query = self.filter(user_id=user.id, idp=idp) + query.update(external_id=external_id, **defaults) + identity_model = query.get() + logger.info( + "updated-identity", + extra={ + "external_id": external_id, + "idp_id": idp.id, + "user_id": user.id, + "identity_id": identity_model.id, + }, + ) + return identity_model + + +@control_silo_model +class Identity(Model): + """ + A verified link between a user and a third party identity. + """ + + __relocation_scope__ = RelocationScope.Excluded + + idp = FlexibleForeignKey("sentry.IdentityProvider") + user = FlexibleForeignKey(settings.AUTH_USER_MODEL) + external_id = models.TextField() + data: models.Field[dict[str, Any], dict[str, Any]] = JSONField() + status = BoundedPositiveIntegerField(default=IdentityStatus.UNKNOWN) + scopes = ArrayField() + date_verified = models.DateTimeField(default=timezone.now) + date_added = models.DateTimeField(default=timezone.now) + + objects: ClassVar[IdentityManager] = IdentityManager() + + class Meta: + app_label = "sentry" + db_table = "sentry_identity" + unique_together = (("idp", "external_id"), ("idp", "user")) + + def get_provider(self) -> Provider: + from sentry.identity import get + + return get(self.idp.type) diff --git a/src/sentry/users/models/user.py b/src/sentry/users/models/user.py index 96c884f32eeb6a..3243d4d6ae18ec 100644 --- a/src/sentry/users/models/user.py +++ b/src/sentry/users/models/user.py @@ -337,9 +337,9 @@ def merge_to(from_user: User, to_user: User) -> None: # TODO: we could discover relations automatically and make this useful from sentry.models.auditlogentry import AuditLogEntry from sentry.models.authidentity import AuthIdentity - from sentry.models.identity import Identity from sentry.models.organizationmembermapping import OrganizationMemberMapping from sentry.users.models.authenticator import Authenticator + from sentry.users.models.identity import Identity from sentry.users.models.user_avatar import UserAvatar from sentry.users.models.user_option import UserOption from sentry.users.models.useremail import UserEmail diff --git a/src/sentry/web/frontend/account_identity.py b/src/sentry/web/frontend/account_identity.py index 7458ccb6d3e337..02a336836d4611 100644 --- a/src/sentry/web/frontend/account_identity.py +++ b/src/sentry/web/frontend/account_identity.py @@ -5,7 +5,7 @@ from rest_framework.request import Request from sentry.identity.pipeline import IdentityProviderPipeline -from sentry.models.identity import IdentityProvider +from sentry.users.models.identity import IdentityProvider from sentry.web.frontend.base import ControlSiloOrganizationView, control_silo_view from sentry.web.helpers import render_to_response diff --git a/tests/acceptance/test_link_team.py b/tests/acceptance/test_link_team.py index 05ad4713ba4913..dde4af1e0f35ae 100644 --- a/tests/acceptance/test_link_team.py +++ b/tests/acceptance/test_link_team.py @@ -5,9 +5,9 @@ from sentry.integrations.models.external_actor import ExternalActor from sentry.integrations.slack.views.link_team import build_team_linking_url from sentry.integrations.types import ExternalProviders -from sentry.models.identity import Identity, IdentityStatus from sentry.testutils.cases import AcceptanceTestCase from sentry.testutils.silo import no_silo_test +from sentry.users.models.identity import Identity, IdentityStatus @no_silo_test diff --git a/tests/sentry/api/endpoints/test_user_identity.py b/tests/sentry/api/endpoints/test_user_identity.py index 4f8b9faf711db9..78c03461dd1d67 100644 --- a/tests/sentry/api/endpoints/test_user_identity.py +++ b/tests/sentry/api/endpoints/test_user_identity.py @@ -1,6 +1,6 @@ -from sentry.models.identity import Identity, IdentityStatus from sentry.testutils.cases import APITestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityStatus @control_silo_test diff --git a/tests/sentry/api/endpoints/test_user_identity_config.py b/tests/sentry/api/endpoints/test_user_identity_config.py index e9ab09263f2415..7b3f0e2e70b5d4 100644 --- a/tests/sentry/api/endpoints/test_user_identity_config.py +++ b/tests/sentry/api/endpoints/test_user_identity_config.py @@ -2,9 +2,9 @@ from sentry.models.authidentity import AuthIdentity from sentry.models.authprovider import AuthProvider -from sentry.models.identity import Identity from sentry.testutils.cases import APITestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity from social_auth.models import UserSocialAuth diff --git a/tests/sentry/api/serializers/test_user_identity_config.py b/tests/sentry/api/serializers/test_user_identity_config.py index b3210b7620f7dc..fbb3996812a164 100644 --- a/tests/sentry/api/serializers/test_user_identity_config.py +++ b/tests/sentry/api/serializers/test_user_identity_config.py @@ -5,9 +5,9 @@ from sentry.api.serializers.models.user_identity_config import Status, UserIdentityConfig from sentry.models.authidentity import AuthIdentity from sentry.models.authprovider import AuthProvider -from sentry.models.identity import Identity from sentry.testutils.cases import TestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity from social_auth.models import UserSocialAuth diff --git a/tests/sentry/integrations/api/endpoints/test_organization_integration_details.py b/tests/sentry/integrations/api/endpoints/test_organization_integration_details.py index a1819980e9b426..e80a2285ce15a3 100644 --- a/tests/sentry/integrations/api/endpoints/test_organization_integration_details.py +++ b/tests/sentry/integrations/api/endpoints/test_organization_integration_details.py @@ -3,13 +3,13 @@ from sentry.integrations.base import IntegrationInstallation from sentry.integrations.models.integration import Integration from sentry.integrations.models.organization_integration import OrganizationIntegration -from sentry.models.identity import Identity from sentry.models.repository import Repository from sentry.models.scheduledeletion import ScheduledDeletion from sentry.shared_integrations.exceptions import ApiError, IntegrationError from sentry.silo.base import SiloMode from sentry.testutils.cases import APITestCase from sentry.testutils.silo import assume_test_silo_mode, control_silo_test +from sentry.users.models.identity import Identity class OrganizationIntegrationDetailsTest(APITestCase): diff --git a/tests/sentry/integrations/bitbucket_server/test_integration.py b/tests/sentry/integrations/bitbucket_server/test_integration.py index 02cf38dcf4dcd8..be5ad281634316 100644 --- a/tests/sentry/integrations/bitbucket_server/test_integration.py +++ b/tests/sentry/integrations/bitbucket_server/test_integration.py @@ -5,9 +5,9 @@ from sentry.integrations.bitbucket_server import BitbucketServerIntegrationProvider from sentry.integrations.models.integration import Integration from sentry.integrations.models.organization_integration import OrganizationIntegration -from sentry.models.identity import Identity, IdentityProvider from sentry.testutils.cases import IntegrationTestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityProvider @control_silo_test diff --git a/tests/sentry/integrations/bitbucket_server/test_repository.py b/tests/sentry/integrations/bitbucket_server/test_repository.py index 1cb43476b9bc77..fe6c7708da8206 100644 --- a/tests/sentry/integrations/bitbucket_server/test_repository.py +++ b/tests/sentry/integrations/bitbucket_server/test_repository.py @@ -17,12 +17,12 @@ REPO, ) from sentry.integrations.bitbucket_server.repository import BitbucketServerRepositoryProvider -from sentry.models.identity import Identity, IdentityStatus from sentry.models.repository import Repository from sentry.shared_integrations.exceptions import IntegrationError from sentry.silo.base import SiloMode from sentry.testutils.cases import APITestCase from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity, IdentityStatus class BitbucketServerRepositoryProviderTest(APITestCase): diff --git a/tests/sentry/integrations/bitbucket_server/test_webhook.py b/tests/sentry/integrations/bitbucket_server/test_webhook.py index 9b33f8dc21492c..40f11225558953 100644 --- a/tests/sentry/integrations/bitbucket_server/test_webhook.py +++ b/tests/sentry/integrations/bitbucket_server/test_webhook.py @@ -7,11 +7,11 @@ from fixtures.bitbucket_server import EXAMPLE_PRIVATE_KEY from sentry.integrations.bitbucket_server.webhook import PROVIDER_NAME -from sentry.models.identity import Identity from sentry.models.repository import Repository from sentry.silo.base import SiloMode from sentry.testutils.cases import APITestCase from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity from sentry_plugins.bitbucket.testutils import REFS_CHANGED_EXAMPLE PROVIDER = "bitbucket_server" diff --git a/tests/sentry/integrations/discord/test_views.py b/tests/sentry/integrations/discord/test_views.py index b7a8e8de59f7c1..e273d06bd21222 100644 --- a/tests/sentry/integrations/discord/test_views.py +++ b/tests/sentry/integrations/discord/test_views.py @@ -4,9 +4,9 @@ from sentry.integrations.discord.views.link_identity import build_linking_url from sentry.integrations.discord.views.unlink_identity import build_unlinking_url -from sentry.models.identity import Identity, IdentityStatus from sentry.testutils.cases import TestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityStatus class DiscordIntegrationLinkIdentityTestBase(TestCase): diff --git a/tests/sentry/integrations/github/test_search.py b/tests/sentry/integrations/github/test_search.py index f2eeca7e42f3eb..2ba71f5e9dd056 100644 --- a/tests/sentry/integrations/github/test_search.py +++ b/tests/sentry/integrations/github/test_search.py @@ -4,9 +4,9 @@ from django.urls import reverse from sentry.integrations.models.organization_integration import OrganizationIntegration -from sentry.models.identity import Identity from sentry.testutils.cases import APITestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity @control_silo_test diff --git a/tests/sentry/integrations/github_enterprise/test_integration.py b/tests/sentry/integrations/github_enterprise/test_integration.py index 5c6b6d5bbf980e..06c4a2b54616a2 100644 --- a/tests/sentry/integrations/github_enterprise/test_integration.py +++ b/tests/sentry/integrations/github_enterprise/test_integration.py @@ -14,12 +14,12 @@ FileBlameInfo, SourceLineInfo, ) -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.models.repository import Repository from sentry.silo.base import SiloMode from sentry.testutils.cases import IntegrationTestCase from sentry.testutils.helpers.integrations import get_installation_of_type from sentry.testutils.silo import assume_test_silo_mode, control_silo_test +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus @control_silo_test diff --git a/tests/sentry/integrations/gitlab/test_client.py b/tests/sentry/integrations/gitlab/test_client.py index eec7fae8f8485d..d5cf668a651234 100644 --- a/tests/sentry/integrations/gitlab/test_client.py +++ b/tests/sentry/integrations/gitlab/test_client.py @@ -23,10 +23,10 @@ FileBlameInfo, SourceLineInfo, ) -from sentry.models.identity import Identity from sentry.shared_integrations.exceptions import ApiError, ApiHostError, ApiRateLimitedError from sentry.testutils.outbox import outbox_runner from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity from sentry.utils.cache import cache GITLAB_CODEOWNERS = { diff --git a/tests/sentry/integrations/gitlab/test_integration.py b/tests/sentry/integrations/gitlab/test_integration.py index 4b2ced66c7f60e..b32b6e8fd70301 100644 --- a/tests/sentry/integrations/gitlab/test_integration.py +++ b/tests/sentry/integrations/gitlab/test_integration.py @@ -21,7 +21,6 @@ FileBlameInfo, SourceLineInfo, ) -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.models.repository import Repository from sentry.shared_integrations.exceptions import ApiUnauthorized from sentry.silo.base import SiloMode @@ -29,6 +28,7 @@ from sentry.testutils.cases import IntegrationTestCase from sentry.testutils.helpers.integrations import get_installation_of_type from sentry.testutils.silo import assume_test_silo_mode, control_silo_test +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus from tests.sentry.integrations.test_helpers import add_control_silo_proxy_response diff --git a/tests/sentry/integrations/gitlab/test_repository.py b/tests/sentry/integrations/gitlab/test_repository.py index c712c5fc2fac60..88137f91514821 100644 --- a/tests/sentry/integrations/gitlab/test_repository.py +++ b/tests/sentry/integrations/gitlab/test_repository.py @@ -6,7 +6,6 @@ from fixtures.gitlab import COMMIT_DIFF_RESPONSE, COMMIT_LIST_RESPONSE, COMPARE_RESPONSE from sentry.integrations.gitlab.repository import GitlabRepositoryProvider -from sentry.models.identity import Identity from sentry.models.pullrequest import PullRequest from sentry.models.repository import Repository from sentry.shared_integrations.exceptions import IntegrationError @@ -14,6 +13,7 @@ from sentry.testutils.asserts import assert_commit_shape from sentry.testutils.cases import IntegrationRepositoryTestCase from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity class GitLabRepositoryProviderTest(IntegrationRepositoryTestCase): diff --git a/tests/sentry/integrations/jira_server/__init__.py b/tests/sentry/integrations/jira_server/__init__.py index 261fa8a16da5cb..9d3e8a17f7844c 100644 --- a/tests/sentry/integrations/jira_server/__init__.py +++ b/tests/sentry/integrations/jira_server/__init__.py @@ -6,10 +6,10 @@ from sentry.integrations.models.integration import Integration from sentry.models.group import Group from sentry.models.grouplink import GroupLink -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.models.organization import Organization from sentry.silo.base import SiloMode from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.users.models.user import User EXAMPLE_PRIVATE_KEY = """-----BEGIN RSA PRIVATE KEY----- diff --git a/tests/sentry/integrations/jira_server/test_installation.py b/tests/sentry/integrations/jira_server/test_installation.py index 99d92a3546bee5..487cf5b6b1c4e1 100644 --- a/tests/sentry/integrations/jira_server/test_installation.py +++ b/tests/sentry/integrations/jira_server/test_installation.py @@ -5,9 +5,9 @@ from sentry.integrations.jira_server import JiraServerIntegrationProvider from sentry.integrations.models.integration import Integration from sentry.integrations.models.organization_integration import OrganizationIntegration -from sentry.models.identity import Identity, IdentityProvider from sentry.testutils.cases import IntegrationTestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityProvider from sentry.utils import jwt from . import EXAMPLE_PRIVATE_KEY diff --git a/tests/sentry/integrations/models/deletions/test_organizationintegration.py b/tests/sentry/integrations/models/deletions/test_organizationintegration.py index 6465a79338a86e..5c55cd6ee218d7 100644 --- a/tests/sentry/integrations/models/deletions/test_organizationintegration.py +++ b/tests/sentry/integrations/models/deletions/test_organizationintegration.py @@ -3,7 +3,6 @@ from sentry.integrations.models.integration import Integration from sentry.integrations.models.organization_integration import OrganizationIntegration from sentry.integrations.models.repository_project_path_config import RepositoryProjectPathConfig -from sentry.models.identity import Identity from sentry.models.project import Project from sentry.models.projectcodeowners import ProjectCodeOwners from sentry.models.repository import Repository @@ -14,6 +13,7 @@ from sentry.testutils.hybrid_cloud import HybridCloudTestMixin from sentry.testutils.outbox import outbox_runner from sentry.testutils.silo import assume_test_silo_mode, control_silo_test +from sentry.users.models.identity import Identity @control_silo_test diff --git a/tests/sentry/integrations/msteams/test_action_state_change.py b/tests/sentry/integrations/msteams/test_action_state_change.py index 7bdb0c7715fc99..f41994c0e639a0 100644 --- a/tests/sentry/integrations/msteams/test_action_state_change.py +++ b/tests/sentry/integrations/msteams/test_action_state_change.py @@ -16,12 +16,12 @@ from sentry.models.authprovider import AuthProvider from sentry.models.group import Group, GroupStatus from sentry.models.groupassignee import GroupAssignee -from sentry.models.identity import Identity, IdentityStatus from sentry.silo.base import SiloMode from sentry.testutils.asserts import assert_mock_called_once_with_partial from sentry.testutils.cases import APITestCase from sentry.testutils.silo import assume_test_silo_mode from sentry.testutils.skips import requires_snuba +from sentry.users.models.identity import Identity, IdentityStatus pytestmark = [requires_snuba] diff --git a/tests/sentry/integrations/msteams/test_link_identity.py b/tests/sentry/integrations/msteams/test_link_identity.py index a7cff9a6e1d591..913bdb7a5ee99c 100644 --- a/tests/sentry/integrations/msteams/test_link_identity.py +++ b/tests/sentry/integrations/msteams/test_link_identity.py @@ -5,9 +5,9 @@ import responses from sentry.integrations.msteams.link_identity import build_linking_url -from sentry.models.identity import Identity, IdentityStatus from sentry.testutils.cases import TestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityStatus @control_silo_test diff --git a/tests/sentry/integrations/msteams/test_unlink_identity.py b/tests/sentry/integrations/msteams/test_unlink_identity.py index eb42a0ced3b8d9..8df5edbb05a889 100644 --- a/tests/sentry/integrations/msteams/test_unlink_identity.py +++ b/tests/sentry/integrations/msteams/test_unlink_identity.py @@ -4,9 +4,9 @@ from sentry.integrations.msteams.constants import SALT from sentry.integrations.msteams.unlink_identity import build_unlinking_url -from sentry.models.identity import Identity, IdentityStatus from sentry.testutils.cases import TestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityStatus from sentry.utils.signing import unsign diff --git a/tests/sentry/integrations/msteams/test_webhook.py b/tests/sentry/integrations/msteams/test_webhook.py index b95da9bbc8ac44..e253f860952b70 100644 --- a/tests/sentry/integrations/msteams/test_webhook.py +++ b/tests/sentry/integrations/msteams/test_webhook.py @@ -10,10 +10,10 @@ from sentry.integrations.models.integration import Integration from sentry.integrations.msteams.utils import ACTION_TYPE -from sentry.models.identity import Identity from sentry.silo.base import SiloMode from sentry.testutils.cases import APITestCase from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity from sentry.utils import jwt from .test_helpers import ( diff --git a/tests/sentry/integrations/slack/notifications/test_issue_alert.py b/tests/sentry/integrations/slack/notifications/test_issue_alert.py index 7a98394c4fdd43..171fdf9ba9c996 100644 --- a/tests/sentry/integrations/slack/notifications/test_issue_alert.py +++ b/tests/sentry/integrations/slack/notifications/test_issue_alert.py @@ -16,7 +16,6 @@ from sentry.integrations.types import ExternalProviders from sentry.issues.grouptype import MonitorIncidentType from sentry.issues.issue_occurrence import IssueEvidence, IssueOccurrence -from sentry.models.identity import Identity, IdentityStatus from sentry.models.notificationsettingoption import NotificationSettingOption from sentry.models.notificationsettingprovider import NotificationSettingProvider from sentry.models.projectownership import ProjectOwnership @@ -33,6 +32,7 @@ from sentry.testutils.helpers.notifications import TEST_ISSUE_OCCURRENCE, TEST_PERF_ISSUE_OCCURRENCE from sentry.testutils.silo import assume_test_silo_mode from sentry.testutils.skips import requires_snuba +from sentry.users.models.identity import Identity, IdentityStatus pytestmark = [requires_snuba] diff --git a/tests/sentry/integrations/slack/test_integration.py b/tests/sentry/integrations/slack/test_integration.py index f0d4a00382b27c..f8738c692e2723 100644 --- a/tests/sentry/integrations/slack/test_integration.py +++ b/tests/sentry/integrations/slack/test_integration.py @@ -14,9 +14,9 @@ from sentry.integrations.slack import SlackIntegration, SlackIntegrationProvider from sentry.integrations.slack.utils.users import SLACK_GET_USERS_PAGE_SIZE from sentry.models.auditlogentry import AuditLogEntry -from sentry.models.identity import Identity, IdentityProvider, IdentityStatus from sentry.testutils.cases import APITestCase, IntegrationTestCase, TestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityProvider, IdentityStatus @control_silo_test diff --git a/tests/sentry/integrations/slack/test_link_identity.py b/tests/sentry/integrations/slack/test_link_identity.py index e09724b2a39aa3..d039e032b245a3 100644 --- a/tests/sentry/integrations/slack/test_link_identity.py +++ b/tests/sentry/integrations/slack/test_link_identity.py @@ -7,10 +7,10 @@ from sentry.integrations.slack.views.link_identity import build_linking_url from sentry.integrations.slack.views.unlink_identity import build_unlinking_url -from sentry.models.identity import Identity, IdentityStatus from sentry.testutils.cases import TestCase from sentry.testutils.helpers import add_identity, install_slack from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityStatus class SlackIntegrationLinkIdentityTestBase(TestCase): diff --git a/tests/sentry/integrations/slack/webhooks/actions/test_enable_notifications.py b/tests/sentry/integrations/slack/webhooks/actions/test_enable_notifications.py index 06c8ed675af5c5..66a67218797ebe 100644 --- a/tests/sentry/integrations/slack/webhooks/actions/test_enable_notifications.py +++ b/tests/sentry/integrations/slack/webhooks/actions/test_enable_notifications.py @@ -4,10 +4,10 @@ ENABLE_SLACK_SUCCESS_MESSAGE, NO_IDENTITY_MESSAGE, ) -from sentry.models.identity import Identity from sentry.models.notificationsettingprovider import NotificationSettingProvider from sentry.silo.base import SiloMode from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity from . import BaseEventTest diff --git a/tests/sentry/integrations/slack/webhooks/actions/test_status.py b/tests/sentry/integrations/slack/webhooks/actions/test_status.py index cdfe7815b26ad5..bcb92d27fa2d47 100644 --- a/tests/sentry/integrations/slack/webhooks/actions/test_status.py +++ b/tests/sentry/integrations/slack/webhooks/actions/test_status.py @@ -22,7 +22,6 @@ from sentry.models.groupassignee import GroupAssignee from sentry.models.groupresolution import GroupResolution from sentry.models.groupsnooze import GroupSnooze -from sentry.models.identity import Identity from sentry.models.organizationmember import InviteStatus, OrganizationMember from sentry.models.release import Release from sentry.models.team import Team @@ -34,6 +33,7 @@ from sentry.testutils.silo import assume_test_silo_mode from sentry.testutils.skips import requires_snuba from sentry.types.group import GroupSubStatus +from sentry.users.models.identity import Identity from sentry.utils.http import absolute_uri from sentry.utils.samples import load_data diff --git a/tests/sentry/integrations/slack/webhooks/commands/__init__.py b/tests/sentry/integrations/slack/webhooks/commands/__init__.py index 20224fb2002e26..808a24873610b9 100644 --- a/tests/sentry/integrations/slack/webhooks/commands/__init__.py +++ b/tests/sentry/integrations/slack/webhooks/commands/__init__.py @@ -16,12 +16,12 @@ from sentry import options from sentry.integrations.slack.utils.auth import set_signing_secret from sentry.integrations.types import EXTERNAL_PROVIDERS, ExternalProviders -from sentry.models.identity import Identity from sentry.models.team import Team from sentry.silo.base import SiloMode from sentry.testutils.cases import APITestCase, TestCase from sentry.testutils.helpers import find_identity, install_slack, link_team, link_user from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity class SlackCommandsTest(APITestCase, TestCase): diff --git a/tests/sentry/integrations/slack/webhooks/commands/test_link_user.py b/tests/sentry/integrations/slack/webhooks/commands/test_link_user.py index b98c0ca2790ed6..e1da2d55e0d6ce 100644 --- a/tests/sentry/integrations/slack/webhooks/commands/test_link_user.py +++ b/tests/sentry/integrations/slack/webhooks/commands/test_link_user.py @@ -5,9 +5,9 @@ build_unlinking_url, ) from sentry.integrations.slack.webhooks.base import NOT_LINKED_MESSAGE -from sentry.models.identity import Identity from sentry.testutils.helpers import get_response_text from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity from tests.sentry.integrations.slack.webhooks.commands import SlackCommandsTest diff --git a/tests/sentry/integrations/slack/webhooks/events/test_discover_link_shared.py b/tests/sentry/integrations/slack/webhooks/events/test_discover_link_shared.py index e5acb15dbf1a06..d6f7f335cec3be 100644 --- a/tests/sentry/integrations/slack/webhooks/events/test_discover_link_shared.py +++ b/tests/sentry/integrations/slack/webhooks/events/test_discover_link_shared.py @@ -8,9 +8,9 @@ from slack_sdk.web import SlackResponse from sentry.integrations.slack.unfurl.types import Handler, LinkType, make_type_coercer -from sentry.models.identity import Identity, IdentityStatus from sentry.silo.base import SiloMode from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity, IdentityStatus from . import LINK_SHARED_EVENT, BaseEventTest diff --git a/tests/sentry/integrations/slack/webhooks/events/test_message_im.py b/tests/sentry/integrations/slack/webhooks/events/test_message_im.py index f3a81d2507136c..4b01d47893b28b 100644 --- a/tests/sentry/integrations/slack/webhooks/events/test_message_im.py +++ b/tests/sentry/integrations/slack/webhooks/events/test_message_im.py @@ -5,11 +5,11 @@ import responses from slack_sdk.web import SlackResponse -from sentry.models.identity import Identity, IdentityStatus from sentry.silo.base import SiloMode from sentry.testutils.cases import IntegratedApiTestCase from sentry.testutils.helpers import get_response_text, override_options from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity, IdentityStatus from . import BaseEventTest diff --git a/tests/sentry/integrations/test_pipeline.py b/tests/sentry/integrations/test_pipeline.py index d5e427373573db..a4ea229b80673a 100644 --- a/tests/sentry/integrations/test_pipeline.py +++ b/tests/sentry/integrations/test_pipeline.py @@ -6,7 +6,6 @@ from sentry.integrations.gitlab.integration import GitlabIntegrationProvider from sentry.integrations.models.integration import Integration from sentry.integrations.models.organization_integration import OrganizationIntegration -from sentry.models.identity import Identity from sentry.models.organizationmapping import OrganizationMapping from sentry.models.repository import Repository from sentry.organizations.absolute_url import generate_organization_url @@ -18,6 +17,7 @@ from sentry.testutils.cases import IntegrationTestCase from sentry.testutils.outbox import outbox_runner from sentry.testutils.silo import assume_test_silo_mode, control_silo_test +from sentry.users.models.identity import Identity class ExamplePlugin(IssuePlugin2): diff --git a/tests/sentry/integrations/vsts/test_client.py b/tests/sentry/integrations/vsts/test_client.py index 68f820ccd3c4e7..571b7a50bfca65 100644 --- a/tests/sentry/integrations/vsts/test_client.py +++ b/tests/sentry/integrations/vsts/test_client.py @@ -15,13 +15,13 @@ from fixtures.vsts import VstsIntegrationTestCase from sentry.integrations.vsts.client import VstsApiClient from sentry.integrations.vsts.integration import VstsIntegration, VstsIntegrationProvider -from sentry.models.identity import Identity, IdentityProvider from sentry.models.repository import Repository from sentry.shared_integrations.exceptions import ApiError from sentry.silo.base import SiloMode from sentry.silo.util import PROXY_BASE_PATH, PROXY_OI_HEADER, PROXY_PATH, PROXY_SIGNATURE_HEADER from sentry.testutils.helpers.integrations import get_installation_of_type from sentry.testutils.silo import assume_test_silo_mode, control_silo_test +from sentry.users.models.identity import Identity, IdentityProvider @control_silo_test diff --git a/tests/sentry/integrations/vsts/test_integration.py b/tests/sentry/integrations/vsts/test_integration.py index de86e0a0f788fd..4059c5545d188e 100644 --- a/tests/sentry/integrations/vsts/test_integration.py +++ b/tests/sentry/integrations/vsts/test_integration.py @@ -12,11 +12,11 @@ from sentry.integrations.models.integration_external_project import IntegrationExternalProject from sentry.integrations.models.organization_integration import OrganizationIntegration from sentry.integrations.vsts import VstsIntegration, VstsIntegrationProvider -from sentry.models.identity import Identity from sentry.models.repository import Repository from sentry.shared_integrations.exceptions import IntegrationError, IntegrationProviderError from sentry.silo.base import SiloMode from sentry.testutils.silo import assume_test_silo_mode, control_silo_test +from sentry.users.models.identity import Identity FULL_SCOPES = ["vso.code", "vso.graph", "vso.serviceendpoint_manage", "vso.work_write"] LIMITED_SCOPES = ["vso.graph", "vso.serviceendpoint_manage", "vso.work_write"] diff --git a/tests/sentry/integrations/vsts/test_issues.py b/tests/sentry/integrations/vsts/test_issues.py index 5c53e6d5579fa3..3715183a16414e 100644 --- a/tests/sentry/integrations/vsts/test_issues.py +++ b/tests/sentry/integrations/vsts/test_issues.py @@ -22,7 +22,6 @@ from sentry.integrations.models.integration_external_project import IntegrationExternalProject from sentry.integrations.services.integration import integration_service from sentry.integrations.vsts.integration import VstsIntegration -from sentry.models.identity import Identity from sentry.shared_integrations.exceptions import IntegrationError from sentry.silo.base import SiloMode from sentry.silo.util import PROXY_PATH @@ -30,6 +29,7 @@ from sentry.testutils.helpers.datetime import before_now, iso_format from sentry.testutils.silo import assume_test_silo_mode, region_silo_test from sentry.testutils.skips import requires_snuba +from sentry.users.models.identity import Identity from sentry.users.services.user.service import user_service pytestmark = [requires_snuba] diff --git a/tests/sentry/integrations/vsts/test_kickoff_vsts_subscription_check.py b/tests/sentry/integrations/vsts/test_kickoff_vsts_subscription_check.py index d192caf8252472..4e89f98b44d7b9 100644 --- a/tests/sentry/integrations/vsts/test_kickoff_vsts_subscription_check.py +++ b/tests/sentry/integrations/vsts/test_kickoff_vsts_subscription_check.py @@ -8,9 +8,9 @@ from sentry.integrations.vsts.tasks.kickoff_subscription_check import ( kickoff_vsts_subscription_check, ) -from sentry.models.identity import Identity from sentry.testutils.cases import TestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity PROVIDER = "vsts" diff --git a/tests/sentry/integrations/vsts/test_provider.py b/tests/sentry/integrations/vsts/test_provider.py index c5481093a26bc8..d733781b892710 100644 --- a/tests/sentry/integrations/vsts/test_provider.py +++ b/tests/sentry/integrations/vsts/test_provider.py @@ -12,9 +12,9 @@ from sentry.identity.vsts.provider import VSTSIdentityProvider, VSTSOAuth2CallbackView from sentry.integrations.vsts.integration import AccountConfigView, AccountForm -from sentry.models.identity import Identity from sentry.testutils.cases import TestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity @control_silo_test diff --git a/tests/sentry/integrations/vsts/test_repository.py b/tests/sentry/integrations/vsts/test_repository.py index f3c961b48fedf0..534ffca9bc4bf5 100644 --- a/tests/sentry/integrations/vsts/test_repository.py +++ b/tests/sentry/integrations/vsts/test_repository.py @@ -7,11 +7,11 @@ from fixtures.vsts import COMMIT_DETAILS_EXAMPLE, COMPARE_COMMITS_EXAMPLE, FILE_CHANGES_EXAMPLE from sentry.integrations.vsts.repository import VstsRepositoryProvider -from sentry.models.identity import Identity from sentry.models.repository import Repository from sentry.silo.base import SiloMode from sentry.testutils.cases import IntegrationRepositoryTestCase, TestCase from sentry.testutils.silo import assume_test_silo_mode, control_silo_test +from sentry.users.models.identity import Identity @control_silo_test diff --git a/tests/sentry/integrations/vsts/test_webhooks.py b/tests/sentry/integrations/vsts/test_webhooks.py index 15c010fdd7b717..ba68ff4cba0faa 100644 --- a/tests/sentry/integrations/vsts/test_webhooks.py +++ b/tests/sentry/integrations/vsts/test_webhooks.py @@ -17,10 +17,10 @@ from sentry.models.activity import Activity from sentry.models.group import Group, GroupStatus from sentry.models.grouplink import GroupLink -from sentry.models.identity import Identity from sentry.silo.base import SiloMode from sentry.testutils.cases import APITestCase from sentry.testutils.silo import assume_test_silo_mode +from sentry.users.models.identity import Identity from sentry.utils.http import absolute_uri diff --git a/tests/sentry/notifications/test_notifications.py b/tests/sentry/notifications/test_notifications.py index 45ac9b9586084c..6f7af8db2f91b6 100644 --- a/tests/sentry/notifications/test_notifications.py +++ b/tests/sentry/notifications/test_notifications.py @@ -17,7 +17,6 @@ from sentry.models.activity import Activity from sentry.models.group import Group, GroupStatus from sentry.models.groupassignee import GroupAssignee -from sentry.models.identity import Identity, IdentityStatus from sentry.models.notificationsettingoption import NotificationSettingOption from sentry.models.rule import Rule from sentry.notifications.notifications.activity.assigned import AssignedActivityNotification @@ -30,6 +29,7 @@ from sentry.testutils.silo import assume_test_silo_mode, control_silo_test from sentry.testutils.skips import requires_snuba from sentry.types.activity import ActivityType +from sentry.users.models.identity import Identity, IdentityStatus from sentry.users.models.user_option import UserOption from sentry.utils import json diff --git a/tests/sentry/models/test_identity.py b/tests/sentry/users/models/test_identity.py similarity index 100% rename from tests/sentry/models/test_identity.py rename to tests/sentry/users/models/test_identity.py diff --git a/tests/sentry/web/frontend/test_account_identity.py b/tests/sentry/web/frontend/test_account_identity.py index d9ae5cd84afca6..4da1c73aefe0e2 100644 --- a/tests/sentry/web/frontend/test_account_identity.py +++ b/tests/sentry/web/frontend/test_account_identity.py @@ -3,9 +3,9 @@ from sentry import identity from sentry.identity.providers.dummy import DummyProvider -from sentry.models.identity import Identity, IdentityStatus from sentry.testutils.cases import TestCase from sentry.testutils.silo import control_silo_test +from sentry.users.models.identity import Identity, IdentityStatus @control_silo_test
333a39f8da3550effb2c18bffcf92b6b6a5976d1
2020-01-24 05:17:38
Lyn Nagara
ref: Merge Event and EventCommon models (#16563)
false
Merge Event and EventCommon models (#16563)
ref
diff --git a/src/sentry/api/serializers/models/project.py b/src/sentry/api/serializers/models/project.py index 2b883dfef34854..411e33949d4c30 100644 --- a/src/sentry/api/serializers/models/project.py +++ b/src/sentry/api/serializers/models/project.py @@ -17,6 +17,7 @@ from sentry.auth.superuser import is_active_superuser from sentry.constants import StatsPeriod from sentry.digests import backend as digests +from sentry.eventstore.models import DEFAULT_SUBJECT_TEMPLATE from sentry.lang.native.utils import convert_crashreport_count from sentry.models import ( EnvironmentProject, @@ -30,7 +31,6 @@ Release, UserOption, UserReport, - DEFAULT_SUBJECT_TEMPLATE, ) from sentry.utils.data_filters import FilterTypes diff --git a/src/sentry/eventstore/models.py b/src/sentry/eventstore/models.py index ac7736cdd88f8e..78f3f25f05bfc5 100644 --- a/src/sentry/eventstore/models.py +++ b/src/sentry/eventstore/models.py @@ -1,20 +1,33 @@ from __future__ import absolute_import import pytz +import six +import string +from collections import OrderedDict from datetime import datetime from dateutil.parser import parse as parse_date +from django.conf import settings +from django.utils.encoding import force_text +from hashlib import md5 -from sentry.models import EventCommon, EventDict +from sentry import eventtypes +from sentry.interfaces.base import get_interfaces +from sentry.models import EventDict from sentry.db.models import NodeData from sentry.snuba.events import Columns +from sentry.utils import json +from sentry.utils.cache import memoize +from sentry.utils.canonical import CanonicalKeyView +from sentry.utils.safe import get_path, trim +from sentry.utils.strings import truncatechars def ref_func(x): return x.project_id or x.project.id -class Event(EventCommon): +class Event(object): """ Event backed by nodestore and Snuba. """ @@ -25,7 +38,6 @@ def __init__(self, project_id, event_id, group_id=None, data=None, snuba_data=No self.group_id = group_id self.data = data self._snuba_data = snuba_data or {} - super(Event, self).__init__() def __getstate__(self): state = self.__dict__.copy() @@ -107,16 +119,11 @@ def timestamp(self): return self._snuba_data[column] return self.datetime.isoformat() - # ============================================ - # Snuba-only implementations of properties that - # would otherwise require nodestore data. - # ============================================ @property def tags(self): """ - Override of tags property that uses tags from snuba rather than - the nodestore event body. This might be useful for implementing - tag deletions without having to rewrite nodestore blobs. + Tags property uses tags from snuba if loaded otherwise falls back to + nodestore. """ tags_key_column = self.__get_column_name(Columns.TAGS_KEY) tags_value_column = self.__get_column_name(Columns.TAGS_VALUE) @@ -128,10 +135,55 @@ def tags(self): return sorted(zip(keys, values)) else: return [] - else: - return super(Event, self).tags + # Nodestore implementation + try: + rv = sorted( + [ + (t, v) + for t, v in get_path(self.data, "tags", filter=True) or () + if t is not None and v is not None + ] + ) + return rv + except ValueError: + # at one point Sentry allowed invalid tag sets such as (foo, bar) + # vs ((tag, foo), (tag, bar)) + return [] + + def get_tag(self, key): + for t, v in self.tags: + if t == key: + return v + return None + + @property + def release(self): + return self.get_tag("sentry:release") + + @property + def dist(self): + return self.get_tag("sentry:dist") + + @property + def transaction(self): + return self.get_tag("transaction") + + def get_environment(self): + from sentry.models import Environment + + if not hasattr(self, "_environment_cache"): + self._environment_cache = Environment.objects.get( + organization_id=self.project.organization_id, + name=Environment.get_name_or_default(self.get_tag("environment")), + ) + + return self._environment_cache def get_minimal_user(self): + """ + A minimal 'User' interface object that gives us enough information + to render a user badge. + """ from sentry.interfaces.user import User user_id_column = self.__get_column_name(Columns.USER_ID) @@ -157,45 +209,320 @@ def get_minimal_user(self): {"id": user_id, "email": email, "username": username, "ip_address": ip_address} ) - return super(Event, self).get_minimal_user() + return self.get_interface("user") - # If the data for these is available from snuba, we assume - # it was already normalized on the way in and we can just return - # it, otherwise we defer to EventCommon implementation. def get_event_type(self): + """ + Return the type of this event. + + See ``sentry.eventtypes``. + """ column = self.__get_column_name(Columns.TYPE) if column in self._snuba_data: return self._snuba_data[column] - return super(Event, self).get_event_type() + return self.data.get("type", "default") @property def ip_address(self): column = self.__get_column_name(Columns.USER_IP_ADDRESS) if column in self._snuba_data: return self._snuba_data[column] - return super(Event, self).ip_address + + ip_address = get_path(self.data, "user", "ip_address") + if ip_address: + return ip_address + + remote_addr = get_path(self.data, "request", "env", "REMOTE_ADDR") + if remote_addr: + return remote_addr + + return None @property def title(self): column = self.__get_column_name(Columns.TITLE) if column in self._snuba_data: return self._snuba_data[column] - return super(Event, self).title + + et = eventtypes.get(self.get_event_type())() + return et.get_title(self.get_event_metadata()) @property def culprit(self): column = self.__get_column_name(Columns.CULPRIT) if column in self._snuba_data: return self._snuba_data[column] - return super(Event, self).culprit + return self.data.get("culprit") @property def location(self): column = self.__get_column_name(Columns.LOCATION) if column in self._snuba_data: return self._snuba_data[column] - return super(Event, self).location + et = eventtypes.get(self.get_event_type())() + return et.get_location(self.get_event_metadata()) + + @classmethod + def generate_node_id(cls, project_id, event_id): + """ + Returns a deterministic node_id for this event based on the project_id + and event_id which together are globally unique. The event body should + be saved under this key in nodestore so it can be retrieved using the + same generated id when we only have project_id and event_id. + """ + return md5("{}:{}".format(project_id, event_id)).hexdigest() + + # TODO We need a better way to cache these properties. functools32 + # doesn't quite do the trick as there is a reference bug with unsaved + # models. But the current _group_cache thing is also clunky because these + # properties need to be stripped out in __getstate__. + @property + def group(self): + from sentry.models import Group + + if not self.group_id: + return None + if not hasattr(self, "_group_cache"): + self._group_cache = Group.objects.get(id=self.group_id) + return self._group_cache + + @group.setter + def group(self, group): + self.group_id = group.id + self._group_cache = group + + @property + def project(self): + from sentry.models import Project + + if not hasattr(self, "_project_cache"): + self._project_cache = Project.objects.get(id=self.project_id) + return self._project_cache + + @project.setter + def project(self, project): + if project is None: + self.project_id = None + else: + self.project_id = project.id + self._project_cache = project + + def get_interfaces(self): + return CanonicalKeyView(get_interfaces(self.data)) + + @memoize + def interfaces(self): + return self.get_interfaces() + + def get_interface(self, name): + return self.interfaces.get(name) + + def get_event_metadata(self): + """ + Return the metadata of this event. + + See ``sentry.eventtypes``. + """ + # For some inexplicable reason we have some cases where the data + # is completely empty. In that case we want to hobble along + # further. + return self.data.get("metadata") or {} + + def get_grouping_config(self): + """Returns the event grouping config.""" + from sentry.grouping.api import get_grouping_config_dict_for_event_data + + return get_grouping_config_dict_for_event_data(self.data, self.project) + + def get_hashes(self, force_config=None): + """ + Returns the calculated hashes for the event. This uses the stored + information if available. Grouping hashes will take into account + fingerprinting and checksums. + """ + # If we have hashes stored in the data we use them, otherwise we + # fall back to generating new ones from the data. We can only use + # this if we do not force a different config. + if force_config is None: + hashes = self.data.get("hashes") + if hashes is not None: + return hashes + + return filter( + None, [x.get_hash() for x in self.get_grouping_variants(force_config).values()] + ) + + def get_grouping_variants(self, force_config=None, normalize_stacktraces=False): + """ + This is similar to `get_hashes` but will instead return the + grouping components for each variant in a dictionary. + + If `normalize_stacktraces` is set to `True` then the event data will be + modified for `in_app` in addition to event variants being created. This + means that after calling that function the event data has been modified + in place. + """ + from sentry.grouping.api import get_grouping_variants_for_event, load_grouping_config + from sentry.stacktraces.processing import normalize_stacktraces_for_grouping + + # Forcing configs has two separate modes. One is where just the + # config ID is given in which case it's merged with the stored or + # default config dictionary + if force_config is not None: + if isinstance(force_config, six.string_types): + stored_config = self.get_grouping_config() + config = dict(stored_config) + config["id"] = force_config + else: + config = force_config + + # Otherwise we just use the same grouping config as stored. if + # this is None the `get_grouping_variants_for_event` will fill in + # the default. + else: + config = self.data.get("grouping_config") + + config = load_grouping_config(config) + if normalize_stacktraces: + normalize_stacktraces_for_grouping(self.data, config) + + return get_grouping_variants_for_event(self, config) + + def get_primary_hash(self): + # TODO: This *might* need to be protected from an IndexError? + return self.get_hashes()[0] + + @property + def real_message(self): + # XXX(mitsuhiko): this is a transitional attribute that should be + # removed. `message` will be renamed to `search_message` and this + # will become `message`. + return ( + get_path(self.data, "logentry", "formatted") + or get_path(self.data, "logentry", "message") + or "" + ) + + @property + def organization(self): + return self.project.organization + + @property + def version(self): + return self.data.get("version", "5") + + def get_raw_data(self): + """Returns the internal raw event data dict.""" + return dict(self.data.items()) + + @property + def size(self): + return len(json.dumps(dict(self.data))) + + def get_email_subject(self): + template = self.project.get_option("mail:subject_template") + if template: + template = EventSubjectTemplate(template) + else: + template = DEFAULT_SUBJECT_TEMPLATE + return truncatechars(template.safe_substitute(EventSubjectTemplateData(self)), 128).encode( + "utf-8" + ) + + def as_dict(self): + """Returns the data in normalized form for external consumers.""" + # We use a OrderedDict to keep elements ordered for a potential JSON serializer + data = OrderedDict() + data["event_id"] = self.event_id + data["project"] = self.project_id + data["release"] = self.release + data["dist"] = self.dist + data["platform"] = self.platform + data["message"] = self.real_message + data["datetime"] = self.datetime + data["tags"] = [(k.split("sentry:", 1)[-1], v) for (k, v) in self.tags] + for k, v in sorted(six.iteritems(self.data)): + if k in data: + continue + if k == "sdk": + v = {v_k: v_v for v_k, v_v in six.iteritems(v) if v_k != "client_ip"} + data[k] = v + + # for a long time culprit was not persisted. In those cases put + # the culprit in from the group. + if data.get("culprit") is None and self.group_id: + data["culprit"] = self.group.culprit + + # Override title and location with dynamically generated data + data["title"] = self.title + data["location"] = self.location + + return data + + @property + def search_message(self): + """ + The internal search_message attribute is only used for search purposes. + It adds a bunch of data from the metadata and the culprit. + """ + data = self.data + culprit = self.culprit + + event_metadata = self.get_event_metadata() + + if event_metadata is None: + event_metadata = eventtypes.get(self.get_event_type())().get_metadata(self.data) + + message = "" + + if data.get("logentry"): + message += data["logentry"].get("formatted") or data["logentry"].get("message") or "" + + if event_metadata: + for value in six.itervalues(event_metadata): + value_u = force_text(value, errors="replace") + if value_u not in message: + message = u"{} {}".format(message, value_u) + + if culprit and culprit not in message: + culprit_u = force_text(culprit, errors="replace") + message = u"{} {}".format(message, culprit_u) + + return trim(message.strip(), settings.SENTRY_MAX_MESSAGE_LENGTH) def __get_column_name(self, column): # Events are currently populated from the Events dataset return column.value.event_name + + +class EventSubjectTemplate(string.Template): + idpattern = r"(tag:)?[_a-z][_a-z0-9]*" + + +class EventSubjectTemplateData(object): + tag_aliases = {"release": "sentry:release", "dist": "sentry:dist", "user": "sentry:user"} + + def __init__(self, event): + self.event = event + + def __getitem__(self, name): + if name.startswith("tag:"): + name = name[4:] + value = self.event.get_tag(self.tag_aliases.get(name, name)) + if value is None: + raise KeyError + return six.text_type(value) + elif name == "project": + return self.event.project.get_full_name() + elif name == "projectID": + return self.event.project.slug + elif name == "shortID" and self.event.group_id: + return self.event.group.qualified_short_id + elif name == "orgID": + return self.event.organization.slug + elif name == "title": + return self.event.title + raise KeyError + + +DEFAULT_SUBJECT_TEMPLATE = EventSubjectTemplate("$shortID - $title") diff --git a/src/sentry/models/event.py b/src/sentry/models/event.py index 3c64e974ba2724..db49bb9ef34bdd 100644 --- a/src/sentry/models/event.py +++ b/src/sentry/models/event.py @@ -1,23 +1,9 @@ from __future__ import absolute_import -import six -import string - -from collections import OrderedDict -from django.conf import settings -from django.utils.encoding import force_text -from hashlib import md5 - from semaphore.processing import StoreNormalizer -from sentry import eventtypes from sentry.db.models import NodeData -from sentry.interfaces.base import get_interfaces -from sentry.utils import json -from sentry.utils.cache import memoize -from sentry.utils.canonical import CanonicalKeyDict, CanonicalKeyView -from sentry.utils.safe import get_path, trim -from sentry.utils.strings import truncatechars +from sentry.utils.canonical import CanonicalKeyDict class EventDict(CanonicalKeyDict): @@ -40,359 +26,3 @@ def __init__(self, data, skip_renormalization=False, **kwargs): data = normalizer.normalize_event(dict(data)) CanonicalKeyDict.__init__(self, data, **kwargs) - - -class EventCommon(object): - @classmethod - def generate_node_id(cls, project_id, event_id): - """ - Returns a deterministic node_id for this event based on the project_id - and event_id which together are globally unique. The event body should - be saved under this key in nodestore so it can be retrieved using the - same generated id when we only have project_id and event_id. - """ - return md5("{}:{}".format(project_id, event_id)).hexdigest() - - # TODO (alex) We need a better way to cache these properties. functools32 - # doesn't quite do the trick as there is a reference bug with unsaved - # models. But the current _group_cache thing is also clunky because these - # properties need to be stripped out in __getstate__. - @property - def group(self): - from sentry.models import Group - - if not self.group_id: - return None - if not hasattr(self, "_group_cache"): - self._group_cache = Group.objects.get(id=self.group_id) - return self._group_cache - - @group.setter - def group(self, group): - # guard against None to not fail on AttributeError - # otherwise Django 1.10 will swallow it in db.models.base init, but - # consequently fail to remove from kwargs, and you'll get the red herring - # TypeError: 'group' is an invalid keyword argument for this function. - if group is not None: - self.group_id = group.id - self._group_cache = group - - @property - def project(self): - from sentry.models import Project - - if not hasattr(self, "_project_cache"): - self._project_cache = Project.objects.get(id=self.project_id) - return self._project_cache - - @project.setter - def project(self, project): - if project is None: - self.project_id = None - else: - self.project_id = project.id - self._project_cache = project - - def get_interfaces(self): - return CanonicalKeyView(get_interfaces(self.data)) - - @memoize - def interfaces(self): - return self.get_interfaces() - - def get_interface(self, name): - return self.interfaces.get(name) - - def get_event_type(self): - """ - Return the type of this event. - - See ``sentry.eventtypes``. - """ - return self.data.get("type", "default") - - def get_event_metadata(self): - """ - Return the metadata of this event. - - See ``sentry.eventtypes``. - """ - # For some inexplicable reason we have some cases where the data - # is completely empty. In that case we want to hobble along - # further. - return self.data.get("metadata") or {} - - def get_grouping_config(self): - """Returns the event grouping config.""" - from sentry.grouping.api import get_grouping_config_dict_for_event_data - - return get_grouping_config_dict_for_event_data(self.data, self.project) - - def get_hashes(self, force_config=None): - """ - Returns the calculated hashes for the event. This uses the stored - information if available. Grouping hashes will take into account - fingerprinting and checksums. - """ - # If we have hashes stored in the data we use them, otherwise we - # fall back to generating new ones from the data. We can only use - # this if we do not force a different config. - if force_config is None: - hashes = self.data.get("hashes") - if hashes is not None: - return hashes - - return filter( - None, [x.get_hash() for x in self.get_grouping_variants(force_config).values()] - ) - - def get_grouping_variants(self, force_config=None, normalize_stacktraces=False): - """ - This is similar to `get_hashes` but will instead return the - grouping components for each variant in a dictionary. - - If `normalize_stacktraces` is set to `True` then the event data will be - modified for `in_app` in addition to event variants being created. This - means that after calling that function the event data has been modified - in place. - """ - from sentry.grouping.api import get_grouping_variants_for_event, load_grouping_config - from sentry.stacktraces.processing import normalize_stacktraces_for_grouping - - # Forcing configs has two separate modes. One is where just the - # config ID is given in which case it's merged with the stored or - # default config dictionary - if force_config is not None: - if isinstance(force_config, six.string_types): - stored_config = self.get_grouping_config() - config = dict(stored_config) - config["id"] = force_config - else: - config = force_config - - # Otherwise we just use the same grouping config as stored. if - # this is None the `get_grouping_variants_for_event` will fill in - # the default. - else: - config = self.data.get("grouping_config") - - config = load_grouping_config(config) - if normalize_stacktraces: - normalize_stacktraces_for_grouping(self.data, config) - - return get_grouping_variants_for_event(self, config) - - def get_primary_hash(self): - # TODO: This *might* need to be protected from an IndexError? - return self.get_hashes()[0] - - @property - def title(self): - # also see event_manager.py which inserts this for snuba - et = eventtypes.get(self.get_event_type())() - return et.get_title(self.get_event_metadata()) - - @property - def culprit(self): - return self.data.get("culprit") - - @property - def location(self): - # also see event_manager.py which inserts this for snuba - et = eventtypes.get(self.get_event_type())() - return et.get_location(self.get_event_metadata()) - - @property - def real_message(self): - # XXX(mitsuhiko): this is a transitional attribute that should be - # removed. `message` will be renamed to `search_message` and this - # will become `message`. - return ( - get_path(self.data, "logentry", "formatted") - or get_path(self.data, "logentry", "message") - or "" - ) - - @property - def organization(self): - return self.project.organization - - @property - def version(self): - return self.data.get("version", "5") - - @property - def ip_address(self): - ip_address = get_path(self.data, "user", "ip_address") - if ip_address: - return ip_address - - remote_addr = get_path(self.data, "request", "env", "REMOTE_ADDR") - if remote_addr: - return remote_addr - - return None - - @property - def tags(self): - try: - rv = sorted( - [ - (t, v) - for t, v in get_path(self.data, "tags", filter=True) or () - if t is not None and v is not None - ] - ) - return rv - except ValueError: - # at one point Sentry allowed invalid tag sets such as (foo, bar) - # vs ((tag, foo), (tag, bar)) - return [] - - def get_tag(self, key): - for t, v in self.tags: - if t == key: - return v - return None - - @property - def release(self): - return self.get_tag("sentry:release") - - @property - def dist(self): - return self.get_tag("sentry:dist") - - def get_raw_data(self): - """Returns the internal raw event data dict.""" - return dict(self.data.items()) - - @property - def size(self): - return len(json.dumps(dict(self.data))) - - @property - def transaction(self): - return self.get_tag("transaction") - - def get_email_subject(self): - template = self.project.get_option("mail:subject_template") - if template: - template = EventSubjectTemplate(template) - else: - template = DEFAULT_SUBJECT_TEMPLATE - return truncatechars(template.safe_substitute(EventSubjectTemplateData(self)), 128).encode( - "utf-8" - ) - - def get_environment(self): - from sentry.models import Environment - - if not hasattr(self, "_environment_cache"): - self._environment_cache = Environment.objects.get( - organization_id=self.project.organization_id, - name=Environment.get_name_or_default(self.get_tag("environment")), - ) - - return self._environment_cache - - def get_minimal_user(self): - """ - A minimal 'User' interface object that gives us enough information - to render a user badge. - """ - return self.get_interface("user") - - def as_dict(self): - """Returns the data in normalized form for external consumers.""" - # We use a OrderedDict to keep elements ordered for a potential JSON serializer - data = OrderedDict() - data["event_id"] = self.event_id - data["project"] = self.project_id - data["release"] = self.release - data["dist"] = self.dist - data["platform"] = self.platform - data["message"] = self.real_message - data["datetime"] = self.datetime - data["tags"] = [(k.split("sentry:", 1)[-1], v) for (k, v) in self.tags] - for k, v in sorted(six.iteritems(self.data)): - if k in data: - continue - if k == "sdk": - v = {v_k: v_v for v_k, v_v in six.iteritems(v) if v_k != "client_ip"} - data[k] = v - - # for a long time culprit was not persisted. In those cases put - # the culprit in from the group. - if data.get("culprit") is None and self.group_id: - data["culprit"] = self.group.culprit - - # Override title and location with dynamically generated data - data["title"] = self.title - data["location"] = self.location - - return data - - @property - def search_message(self): - """ - The internal search_message attribute is only used for search purposes. - It adds a bunch of data from the metadata and the culprit. - """ - data = self.data - culprit = self.culprit - - event_metadata = self.get_event_metadata() - - if event_metadata is None: - event_metadata = eventtypes.get(self.get_event_type())().get_metadata(self.data) - - message = "" - - if data.get("logentry"): - message += data["logentry"].get("formatted") or data["logentry"].get("message") or "" - - if event_metadata: - for value in six.itervalues(event_metadata): - value_u = force_text(value, errors="replace") - if value_u not in message: - message = u"{} {}".format(message, value_u) - - if culprit and culprit not in message: - culprit_u = force_text(culprit, errors="replace") - message = u"{} {}".format(message, culprit_u) - - return trim(message.strip(), settings.SENTRY_MAX_MESSAGE_LENGTH) - - -class EventSubjectTemplate(string.Template): - idpattern = r"(tag:)?[_a-z][_a-z0-9]*" - - -class EventSubjectTemplateData(object): - tag_aliases = {"release": "sentry:release", "dist": "sentry:dist", "user": "sentry:user"} - - def __init__(self, event): - self.event = event - - def __getitem__(self, name): - if name.startswith("tag:"): - name = name[4:] - value = self.event.get_tag(self.tag_aliases.get(name, name)) - if value is None: - raise KeyError - return six.text_type(value) - elif name == "project": - return self.event.project.get_full_name() - elif name == "projectID": - return self.event.project.slug - elif name == "shortID" and self.event.group_id: - return self.event.group.qualified_short_id - elif name == "orgID": - return self.event.organization.slug - elif name == "title": - return self.event.title - raise KeyError - - -DEFAULT_SUBJECT_TEMPLATE = EventSubjectTemplate("$shortID - $title")
761c66ca6c2506d147aeb299b21ec0f115487206
2018-07-18 05:35:00
Lyn Nagara
fix(discover): Convert aggregation aliases to underscores (#9084)
false
Convert aggregation aliases to underscores (#9084)
fix
diff --git a/src/sentry/static/sentry/app/views/organizationDiscover/aggregations/utils.jsx b/src/sentry/static/sentry/app/views/organizationDiscover/aggregations/utils.jsx index cc8c2d149d57b4..3a3ba54d2564e1 100644 --- a/src/sentry/static/sentry/app/views/organizationDiscover/aggregations/utils.jsx +++ b/src/sentry/static/sentry/app/views/organizationDiscover/aggregations/utils.jsx @@ -85,7 +85,10 @@ export function getExternal(internal) { if (internal.match(uniqRegex)) { const column = internal.match(uniqRegex)[1]; - return ['uniq', column, `uniq_${column}`]; + const tagMatch = column.match(/^tags\[(.+)]$/); + const alias = tagMatch ? `tags_${tagMatch[1]}` : column; + + return ['uniq', column, `uniq_${alias}`]; } if (internal.match(avgRegex)) { diff --git a/tests/js/spec/views/organizationDiscover/aggregations/aggregation.spec.jsx b/tests/js/spec/views/organizationDiscover/aggregations/aggregation.spec.jsx index 46cd2278a8f70e..5bb76d5fa8e257 100644 --- a/tests/js/spec/views/organizationDiscover/aggregations/aggregation.spec.jsx +++ b/tests/js/spec/views/organizationDiscover/aggregations/aggregation.spec.jsx @@ -8,7 +8,7 @@ describe('Aggregation', function() { it('renders empty, count, topK, uniq and avg', function() { const data = [ {value: [null, null, null], expectedTextValue: 'Add aggregation function...'}, - {value: ['count', null, 'count'], expectedTextValue: 'count'}, + {value: ['count()', null, 'count'], expectedTextValue: 'count'}, { value: ['uniq', 'environment', 'uniq_environment'], expectedTextValue: 'uniq(environment)', @@ -21,6 +21,10 @@ describe('Aggregation', function() { value: ['topK(5)', 'environment', 'topK_5_environment'], expectedTextValue: 'topK(5)(environment)', }, + { + value: ['uniq', 'tags[message]', 'uniq_tags_message'], + expectedTextValue: 'uniq(tags[message])', + }, ]; data.forEach(function(item) {
9ae1b8fd172f1c6165f1a109b6eec4ed8498447a
2021-06-22 15:22:56
Evan Purkhiser
chore(dependabot): Ignore _all_ un-upgraded packages (#26776)
false
Ignore _all_ un-upgraded packages (#26776)
chore
diff --git a/.github/dependabot.yml b/.github/dependabot.yml index fd257a54160588..4ebba25169db78 100644 --- a/.github/dependabot.yml +++ b/.github/dependabot.yml @@ -10,16 +10,77 @@ updates: reviewers: - "@getsentry/owners-js-build" ignore: - # These are packages we're "stuck" on for now - - dependency-name: bootstrap - - dependency-name: jquery + # We ignore everything that hasn't yet been upgrade, this way we will + # only get the _freshest_ of new packages to consider upgrading + - dependency-name: "@babel/core" + - dependency-name: "@babel/plugin-proposal-class-properties" + - dependency-name: "@babel/plugin-proposal-decorators" + - dependency-name: "@babel/plugin-transform-react-jsx" + - dependency-name: "@babel/plugin-transform-react-jsx-source" + - dependency-name: "@babel/plugin-transform-runtime" + - dependency-name: "@babel/preset-env" + - dependency-name: "@babel/preset-react" + - dependency-name: "@babel/preset-typescript" + - dependency-name: "@babel/runtime" + - dependency-name: "@dnd-kit/core" + - dependency-name: "@dnd-kit/sortable" + - dependency-name: "@sentry-internal/global-search" + - dependency-name: "@sentry/release-parser" + - dependency-name: "@sentry/rrweb" + - dependency-name: "@size-limit/preset-small-lib" + - dependency-name: "@types/echarts" - dependency-name: "@types/jquery" - - dependency-name: reflux - - dependency-name: "@types/reflux" - - dependency-name: react-bootstrap - - dependency-name: "@types/react-bootstrap" - - dependency-name: react-router + - dependency-name: "@types/marked" + - dependency-name: "@types/react-mentions" - dependency-name: "@types/react-router" - - dependency-name: react-select - dependency-name: "@types/react-select" - - dependency-name: react-select-legacy + - dependency-name: "@types/reflux" + - dependency-name: "@visual-snapshot/jest" + - dependency-name: "@wojtekmaj/enzyme-adapter-react-17" + - dependency-name: "babel-jest" + - dependency-name: "bootstrap" + - dependency-name: "clean-webpack-plugin" + - dependency-name: "compression-webpack-plugin" + - dependency-name: "copy-text-to-clipboard" + - dependency-name: "core-js" + - dependency-name: "create-react-class" + - dependency-name: "css-loader" + - dependency-name: "css-loader" + - dependency-name: "css-minimizer-webpack-plugin" + - dependency-name: "dotenv-webpack" + - dependency-name: "echarts" + - dependency-name: "echarts-for-react" + - dependency-name: "eslint" + - dependency-name: "eslint-plugin-simple-import-sort" + - dependency-name: "fuse.js" + - dependency-name: "gettext-parser" + - dependency-name: "jest" + - dependency-name: "jest-circus" + - dependency-name: "jest-junit" + - dependency-name: "jquery" + - dependency-name: "less" + - dependency-name: "less-loader" + - dependency-name: "less-loader" + - dependency-name: "marked" + - dependency-name: "mini-css-extract-plugin" + - dependency-name: "mobx" + - dependency-name: "mobx-react" + - dependency-name: "papaparse" + - dependency-name: "prettier" + - dependency-name: "query-string" + - dependency-name: "react-date-range" + - dependency-name: "react-lazyload" + - dependency-name: "react-mentions" + - dependency-name: "react-popper" + - dependency-name: "react-refresh" + - dependency-name: "react-router" + - dependency-name: "react-select" + - dependency-name: "reflux" + - dependency-name: "rrweb" + - dependency-name: "rrweb-player" + - dependency-name: "size-limit" + - dependency-name: "sprintf-js" + - dependency-name: "terser-webpack-plugin" + - dependency-name: "tslib" + - dependency-name: "typescript" + - dependency-name: "u2f-api"
8857e10c6612329f72709ab5cdfa410b836ce629
2025-01-15 22:03:15
Nikki Kapadia
feat(widget-builder): Create blank widget preview for templates (#83509)
false
Create blank widget preview for templates (#83509)
feat
diff --git a/static/app/views/dashboards/widgetBuilder/components/newWidgetBuilder.spec.tsx b/static/app/views/dashboards/widgetBuilder/components/newWidgetBuilder.spec.tsx index 7c15aae748fe88..3d5242ca3cf282 100644 --- a/static/app/views/dashboards/widgetBuilder/components/newWidgetBuilder.spec.tsx +++ b/static/app/views/dashboards/widgetBuilder/components/newWidgetBuilder.spec.tsx @@ -240,4 +240,23 @@ describe('NewWidgetBuiler', function () { expect(await screen.findByText('Select group')).toBeInTheDocument(); expect(await screen.findByText('Add Group')).toBeInTheDocument(); }); + + it('renders empty widget preview when no widget selected from templates', async function () { + render( + <WidgetBuilderV2 + isOpen + onClose={onCloseMock} + dashboard={DashboardFixture([])} + dashboardFilters={{}} + onSave={onSaveMock} + openWidgetTemplates + setOpenWidgetTemplates={jest.fn()} + />, + {router, organization} + ); + + expect(await screen.findByText('Add from Widget Library')).toBeInTheDocument(); + + expect(await screen.findByText('Select a widget to preview')).toBeInTheDocument(); + }); }); diff --git a/static/app/views/dashboards/widgetBuilder/components/newWidgetBuilder.tsx b/static/app/views/dashboards/widgetBuilder/components/newWidgetBuilder.tsx index bf2f7a5ec76fc1..c92bc2308ae4cf 100644 --- a/static/app/views/dashboards/widgetBuilder/components/newWidgetBuilder.tsx +++ b/static/app/views/dashboards/widgetBuilder/components/newWidgetBuilder.tsx @@ -4,6 +4,7 @@ import {css, useTheme} from '@emotion/react'; import styled from '@emotion/styled'; import {AnimatePresence, motion} from 'framer-motion'; import cloneDeep from 'lodash/cloneDeep'; +import omit from 'lodash/omit'; import {t} from 'sentry/locale'; import {space} from 'sentry/styles/space'; @@ -169,6 +170,7 @@ function WidgetBuilderV2({ isDraggable={isPreviewDraggable} isWidgetInvalid={!queryConditionsValid} onDataFetched={handleWidgetDataFetched} + openWidgetTemplates={openWidgetTemplates} /> </DndContext> )} @@ -192,6 +194,7 @@ export function WidgetPreviewContainer({ dragPosition, isDraggable, onDataFetched, + openWidgetTemplates, }: { dashboard: DashboardDetails; dashboardFilters: DashboardFilters; @@ -199,6 +202,7 @@ export function WidgetPreviewContainer({ dragPosition?: WidgetDragPositioning; isDraggable?: boolean; onDataFetched?: (tableData: TableDataWithTitle[]) => void; + openWidgetTemplates?: boolean; }) { const {state} = useWidgetBuilderContext(); const organization = useOrganization(); @@ -209,7 +213,6 @@ export function WidgetPreviewContainer({ useRpc: decodeBoolean, }, }); - const isSmallScreen = useMedia(`(max-width: ${theme.breakpoints.small})`); // if small screen and draggable, enable dragging const isDragEnabled = isSmallScreen && isDraggable; @@ -224,7 +227,7 @@ export function WidgetPreviewContainer({ const draggableStyle: CSSProperties = { transform: isDragEnabled - ? `translate3d(${translate?.x ?? 0}px, ${translate?.y ?? 0}px, 0)` + ? `translate3d(${isDragging ? translate?.x : 0}px, ${isDragging ? translate?.y : 0}px, 0)` : undefined, top: isDragEnabled ? top ?? 0 : undefined, left: isDragEnabled ? left ?? 0 : undefined, @@ -236,10 +239,27 @@ export function WidgetPreviewContainer({ position: isDragEnabled ? 'fixed' : undefined, }; + // check if the state is in the url because the state variable has default values + const hasUrlParams = + Object.keys( + omit(location.query, [ + 'environment', + 'project', + 'release', + 'start', + 'end', + 'statsPeriod', + ]) + ).length > 0; + const getPreviewHeight = () => { if (isDragEnabled) { return DRAGGABLE_PREVIEW_HEIGHT_PX; } + // if none of the widget templates are selected + if (openWidgetTemplates && !hasUrlParams) { + return PREVIEW_HEIGHT_PX; + } if (state.displayType === DisplayType.TABLE) { return 'auto'; } @@ -289,14 +309,25 @@ export function WidgetPreviewContainer({ : undefined, }} > - <WidgetPreview - // While we test out RPC for spans, force a re-render if the spans toggle changes - key={state.dataset === WidgetType.SPANS && useRpc ? 'spans' : 'other'} - dashboardFilters={dashboardFilters} - dashboard={dashboard} - isWidgetInvalid={isWidgetInvalid} - onDataFetched={onDataFetched} - /> + {openWidgetTemplates && !hasUrlParams ? ( + <WidgetPreviewPlaceholder> + <h6 style={{margin: 0}}>{t('Widget Title')}</h6> + <TemplateWidgetPreviewPlaceholder> + <p style={{margin: 0}}>{t('Select a widget to preview')}</p> + </TemplateWidgetPreviewPlaceholder> + </WidgetPreviewPlaceholder> + ) : ( + <WidgetPreview + // While we test out RPC for spans, force a re-render if the spans toggle changes + key={ + state.dataset === WidgetType.SPANS && useRpc ? 'spans' : 'other' + } + dashboardFilters={dashboardFilters} + dashboard={dashboard} + isWidgetInvalid={isWidgetInvalid} + onDataFetched={onDataFetched} + /> + )} </SampleWidgetCard> </DraggableWidgetContainer> </MEPSettingProvider> @@ -420,3 +451,22 @@ const DroppableGrid = styled('div')` bottom: ${space(2)}; left: 0; `; + +const TemplateWidgetPreviewPlaceholder = styled('div')` + display: flex; + flex-direction: column; + align-items: center; + justify-content: center; + width: 100%; + height: 95%; + color: ${p => p.theme.subText}; + font-style: italic; + font-size: ${p => p.theme.fontSizeMedium}; + font-weight: ${p => p.theme.fontWeightNormal}; +`; + +const WidgetPreviewPlaceholder = styled('div')` + width: 100%; + height: 100%; + padding: ${space(2)}; +`; diff --git a/static/app/views/dashboards/widgetBuilder/components/widgetBuilderSlideout.tsx b/static/app/views/dashboards/widgetBuilder/components/widgetBuilderSlideout.tsx index 2f63b183788314..d50f0c64d7620a 100644 --- a/static/app/views/dashboards/widgetBuilder/components/widgetBuilderSlideout.tsx +++ b/static/app/views/dashboards/widgetBuilder/components/widgetBuilderSlideout.tsx @@ -84,7 +84,8 @@ function WidgetBuilderSlideout({ state.displayType !== DisplayType.BIG_NUMBER && state.displayType !== DisplayType.TABLE; - const previewRef = useRef<HTMLDivElement>(null); + const customPreviewRef = useRef<HTMLDivElement>(null); + const templatesPreviewRef = useRef<HTMLDivElement>(null); const isSmallScreen = useMedia(`(max-width: ${theme.breakpoints.small})`); @@ -100,12 +101,17 @@ function WidgetBuilderSlideout({ {threshold: 0} ); - if (previewRef.current) { - observer.observe(previewRef.current); + // need two different refs to account for preview when customizing templates + if (customPreviewRef.current) { + observer.observe(customPreviewRef.current); + } + + if (templatesPreviewRef.current) { + observer.observe(templatesPreviewRef.current); } return () => observer.disconnect(); - }, [setIsPreviewDraggable]); + }, [setIsPreviewDraggable, openWidgetTemplates]); return ( <SlideOverPanel @@ -151,7 +157,7 @@ function WidgetBuilderSlideout({ <Section> <WidgetBuilderTypeSelector error={error} setError={setError} /> </Section> - <div ref={previewRef}> + <div ref={customPreviewRef}> {isSmallScreen && ( <Section> <WidgetPreviewContainer @@ -159,6 +165,7 @@ function WidgetBuilderSlideout({ dashboardFilters={dashboardFilters} isWidgetInvalid={isWidgetInvalid} onDataFetched={onDataFetched} + openWidgetTemplates={openWidgetTemplates} /> </Section> )} @@ -196,7 +203,7 @@ function WidgetBuilderSlideout({ </Fragment> ) : ( <Fragment> - <div ref={previewRef}> + <div ref={templatesPreviewRef}> {isSmallScreen && ( <Section> <WidgetPreviewContainer @@ -204,6 +211,7 @@ function WidgetBuilderSlideout({ dashboardFilters={dashboardFilters} isWidgetInvalid={isWidgetInvalid} onDataFetched={onDataFetched} + openWidgetTemplates={openWidgetTemplates} /> </Section> )} @@ -211,6 +219,7 @@ function WidgetBuilderSlideout({ <WidgetTemplatesList onSave={onSave} setOpenWidgetTemplates={setOpenWidgetTemplates} + setIsPreviewDraggable={setIsPreviewDraggable} /> </Fragment> )} diff --git a/static/app/views/dashboards/widgetBuilder/components/widgetTemplatesList.spec.tsx b/static/app/views/dashboards/widgetBuilder/components/widgetTemplatesList.spec.tsx index c64f2af4c6e5db..1edffc32364d34 100644 --- a/static/app/views/dashboards/widgetBuilder/components/widgetTemplatesList.spec.tsx +++ b/static/app/views/dashboards/widgetBuilder/components/widgetTemplatesList.spec.tsx @@ -55,7 +55,11 @@ describe('WidgetTemplatesList', () => { it('should render the widget templates list', async () => { render( <WidgetBuilderProvider> - <WidgetTemplatesList onSave={onSave} setOpenWidgetTemplates={jest.fn()} /> + <WidgetTemplatesList + onSave={onSave} + setOpenWidgetTemplates={jest.fn()} + setIsPreviewDraggable={jest.fn()} + /> </WidgetBuilderProvider> ); @@ -66,7 +70,11 @@ describe('WidgetTemplatesList', () => { it('should render buttons when the user clicks on a widget template', async () => { render( <WidgetBuilderProvider> - <WidgetTemplatesList onSave={onSave} setOpenWidgetTemplates={jest.fn()} /> + <WidgetTemplatesList + onSave={onSave} + setOpenWidgetTemplates={jest.fn()} + setIsPreviewDraggable={jest.fn()} + /> </WidgetBuilderProvider> ); @@ -83,7 +91,11 @@ describe('WidgetTemplatesList', () => { render( <WidgetBuilderProvider> - <WidgetTemplatesList onSave={onSave} setOpenWidgetTemplates={jest.fn()} /> + <WidgetTemplatesList + onSave={onSave} + setOpenWidgetTemplates={jest.fn()} + setIsPreviewDraggable={jest.fn()} + /> </WidgetBuilderProvider>, {router} ); @@ -107,7 +119,11 @@ describe('WidgetTemplatesList', () => { it('should show error message when the widget fails to save', async () => { render( <WidgetBuilderProvider> - <WidgetTemplatesList onSave={onSave} setOpenWidgetTemplates={jest.fn()} /> + <WidgetTemplatesList + onSave={onSave} + setOpenWidgetTemplates={jest.fn()} + setIsPreviewDraggable={jest.fn()} + /> </WidgetBuilderProvider> ); diff --git a/static/app/views/dashboards/widgetBuilder/components/widgetTemplatesList.tsx b/static/app/views/dashboards/widgetBuilder/components/widgetTemplatesList.tsx index 2e082c9356590a..9a1e140b0381c0 100644 --- a/static/app/views/dashboards/widgetBuilder/components/widgetTemplatesList.tsx +++ b/static/app/views/dashboards/widgetBuilder/components/widgetTemplatesList.tsx @@ -19,10 +19,15 @@ import {getWidgetIcon} from 'sentry/views/dashboards/widgetLibrary/widgetCard'; interface WidgetTemplatesListProps { onSave: ({index, widget}: {index: number; widget: Widget}) => void; + setIsPreviewDraggable: (isPreviewDraggable: boolean) => void; setOpenWidgetTemplates: (openWidgetTemplates: boolean) => void; } -function WidgetTemplatesList({onSave, setOpenWidgetTemplates}: WidgetTemplatesListProps) { +function WidgetTemplatesList({ + onSave, + setOpenWidgetTemplates, + setIsPreviewDraggable, +}: WidgetTemplatesListProps) { const organization = useOrganization(); const [selectedWidget, setSelectedWidget] = useState<number | null>(null); @@ -72,7 +77,14 @@ function WidgetTemplatesList({onSave, setOpenWidgetTemplates}: WidgetTemplatesLi <WidgetDescription>{widget.description}</WidgetDescription> {selectedWidget === index && ( <ButtonsWrapper> - <Button size="sm" onClick={() => setOpenWidgetTemplates(false)}> + <Button + size="sm" + onClick={() => { + setOpenWidgetTemplates(false); + // reset preview when customizing templates + setIsPreviewDraggable(false); + }} + > {t('Customize')} </Button> <Button size="sm" onClick={() => handleSave(widget)}>
5f6a5f8297958612dc0ba790952747f9e5f22f19
2022-08-05 00:09:10
David Wang
fix(trace): Fix doc link for connecting-services (#37369)
false
Fix doc link for connecting-services (#37369)
fix
diff --git a/static/app/components/quickTrace/index.tsx b/static/app/components/quickTrace/index.tsx index 9b8e6789209f3e..5d1f52fec69a10 100644 --- a/static/app/components/quickTrace/index.tsx +++ b/static/app/components/quickTrace/index.tsx @@ -588,7 +588,7 @@ class MissingServiceNode extends Component<MissingServiceProps, MissingServiceSt const docsHref = docPlatform === null || docPlatform === 'javascript' ? 'https://docs.sentry.io/platforms/javascript/performance/connect-services/' - : `https://docs.sentry.io/platforms/${docPlatform}/performance/connecting-services`; + : `https://docs.sentry.io/platforms/${docPlatform}/performance/connect-services`; return ( <Fragment> {connectorSide === 'left' && <TraceConnector />}
b5f4ec28d55b1f297aaf6bb63a94834154065f9e
2023-09-18 12:40:26
Armin Ronacher
feat(minimetrics): faster encode and filter out statsd from s4s (#56337)
false
faster encode and filter out statsd from s4s (#56337)
feat
diff --git a/src/minimetrics/transport.py b/src/minimetrics/transport.py index 77876fe99184eb..35b308294c5653 100644 --- a/src/minimetrics/transport.py +++ b/src/minimetrics/transport.py @@ -1,11 +1,12 @@ import re from functools import partial +from io import BytesIO from typing import Iterable import sentry_sdk from sentry_sdk.envelope import Envelope, Item -from minimetrics.types import FlushedMetric, MetricTagsInternal +from minimetrics.types import FlushedMetric from sentry.utils import metrics @@ -21,30 +22,42 @@ class EncodingError(Exception): class RelayStatsdEncoder: - def encode(self, value: FlushedMetric) -> str: + def _encode(self, value: FlushedMetric, out: BytesIO): + _write = out.write (timestamp, metric_type, metric_name, metric_unit, metric_tags), metric = value metric_name = sanitize_value(metric_name) or "invalid-metric-name" - metric_values = ":".join(str(v) for v in metric.serialize_value()) - serialized_metric_tags = self._get_metric_tags(metric_tags) - metric_tags_prefix = serialized_metric_tags and "|#" or "" - return f"{metric_name}@{metric_unit}:{metric_values}|{metric_type}{metric_tags_prefix}{serialized_metric_tags}|T{timestamp}" - - def encode_multiple(self, values: Iterable[FlushedMetric]) -> str: - return "\n".join(self.encode(value) for value in values) - - def _get_metric_tags(self, tags: MetricTagsInternal) -> str: - if not tags: - return "" - - # We sanitize all the tag keys and tag values. - sanitized_tags = ( - (sanitize_value(tag_key), sanitize_value(tag_value)) for tag_key, tag_value in tags - ) - - # We then convert all tags whose tag key is not empty to the string representation. - return ",".join( - f"{tag_key}:{tag_value}" for tag_key, tag_value in sanitized_tags if tag_key - ) + _write(f"{metric_name}@{metric_unit}".encode()) + + for serialized_value in metric.serialize_value(): + _write(b":") + _write(str(serialized_value).encode("utf-8")) + + _write(f"|{metric_type}".encode("ascii")) + + if metric_tags: + _write(b"|#") + first = True + for tag_key, tag_value in metric_tags: + tag_key = sanitize_value(tag_key) + if not tag_key: + continue + if first: + first = False + else: + _write(b",") + _write(tag_key.encode("utf-8")) + _write(b":") + _write(sanitize_value(tag_value).encode("utf-8")) + + _write(f"|T{timestamp}".encode("ascii")) + + def encode_multiple(self, values: Iterable[FlushedMetric]) -> bytes: + out = BytesIO() + _write = out.write + for value in values: + self._encode(value, out) + _write(b"\n") + return out.getvalue() class MetricEnvelopeTransport: diff --git a/src/sentry/utils/sdk.py b/src/sentry/utils/sdk.py index bf4d9e00d8bb90..e25b002b283e3d 100644 --- a/src/sentry/utils/sdk.py +++ b/src/sentry/utils/sdk.py @@ -405,6 +405,15 @@ def _capture_anything(self, method_name, *args, **kwargs): # install_id = options.get('sentry:install-id') # if install_id: # event.setdefault('tags', {})['install-id'] = install_id + if method_name == "capture_envelope": + args_list = list(args) + envelope = args_list[0] + # Do not forward metrics to s4s + safe_items = [x for x in envelope.items if x.data_category != "statsd"] + if len(safe_items) != len(envelope.items): + relay_envelope = copy.copy(envelope) + relay_envelope.items = safe_items + args = [relay_envelope, *args_list[1:]] getattr(sentry4sentry_transport, method_name)(*args, **kwargs) if sentry_saas_transport and options.get("store.use-relay-dsn-sample-rate") == 1: diff --git a/tests/minimetrics/test_transport.py b/tests/minimetrics/test_transport.py index c88626246ac560..a725f481f9ee29 100644 --- a/tests/minimetrics/test_transport.py +++ b/tests/minimetrics/test_transport.py @@ -1,3 +1,4 @@ +import io from typing import Any from unittest.mock import patch @@ -6,9 +7,14 @@ from minimetrics.types import BucketKey -def test_relay_encoder_with_counter(): +def encode_metric(value): encoder = RelayStatsdEncoder() + out = io.BytesIO() + encoder._encode(value, out) + return out.getvalue().decode("utf-8") + +def test_relay_encoder_with_counter(): bucket_key: BucketKey = ( 1693994400, "c", @@ -22,13 +28,11 @@ def test_relay_encoder_with_counter(): metric = CounterMetric(first=2) flushed_metric = (bucket_key, metric) - result = encoder.encode(flushed_metric) + result = encode_metric(flushed_metric) assert result == "button_click@none:2|c|#browser:Chrome,browser.version:1.0|T1693994400" def test_relay_encoder_with_distribution(): - encoder = RelayStatsdEncoder() - bucket_key: BucketKey = ( 1693994400, "d", @@ -44,7 +48,7 @@ def test_relay_encoder_with_distribution(): metric.add(3.0) flushed_metric = (bucket_key, metric) - result = encoder.encode(flushed_metric) + result = encode_metric(flushed_metric) assert ( result == "execution_time@second:1.0:0.5:3.0|d|#browser:Chrome,browser.version:1.0|T1693994400" @@ -52,8 +56,6 @@ def test_relay_encoder_with_distribution(): def test_relay_encoder_with_set(): - encoder = RelayStatsdEncoder() - bucket_key: BucketKey = ( 1693994400, "s", @@ -69,7 +71,7 @@ def test_relay_encoder_with_set(): metric.add("riccardo") flushed_metric = (bucket_key, metric) - result = encoder.encode(flushed_metric) + result = encode_metric(flushed_metric) pieces = result.split("|") m = pieces[0].split(":") @@ -82,8 +84,6 @@ def test_relay_encoder_with_set(): def test_relay_encoder_with_gauge(): - encoder = RelayStatsdEncoder() - bucket_key: BucketKey = ( 1693994400, "g", @@ -99,7 +99,7 @@ def test_relay_encoder_with_gauge(): metric.add(7.0) flushed_metric = (bucket_key, metric) - result = encoder.encode(flushed_metric) + result = encode_metric(flushed_metric) assert ( result == "startup_time@second:7.0:5.0:10.0:22.0:3|g|#browser:Chrome,browser.version:1.0|T1693994400" @@ -107,8 +107,6 @@ def test_relay_encoder_with_gauge(): def test_relay_encoder_with_invalid_chars(): - encoder = RelayStatsdEncoder() - bucket_key: BucketKey = ( 1693994400, "c", @@ -130,7 +128,7 @@ def test_relay_encoder_with_invalid_chars(): metric = CounterMetric(first=1) flushed_metric = (bucket_key, metric) - result = encoder.encode(flushed_metric) + result = encode_metric(flushed_metric) assert ( result == "bttn_click@second:1|c|#browsername:Chrome,browser.version:1.0,platform:Android,version:|T1693994400" @@ -146,7 +144,7 @@ def test_relay_encoder_with_invalid_chars(): metric = CounterMetric(first=1) flushed_metric = (bucket_key, metric) - assert encoder.encode(flushed_metric) == "invalid-metric-name@second:1|c|T1693994400" + assert encode_metric(flushed_metric) == "invalid-metric-name@second:1|c|T1693994400" def test_relay_encoder_with_multiple_metrics(): @@ -196,14 +194,12 @@ def test_relay_encoder_with_multiple_metrics(): ) metrics: Any = [flushed_metric_1, flushed_metric_2, flushed_metric_3] - result = encoder.encode_multiple(metrics) + result = encoder.encode_multiple(metrics).decode("utf-8") assert result == ( - "startup_time@second:10.0:10.0:10.0:10.0:1|g|#browser:Chrome,browser.version:1.0|T1693994400" - + "\n" - + "button_click@none:1|c|#browser:Chrome,browser.version:1.0|T1693994400" - + "\n" - + "invalid-metric-name@none:1|c|#browser:Chrome,browser.version:1.0|T1693994400" + "startup_time@second:10.0:10.0:10.0:10.0:1|g|#browser:Chrome,browser.version:1.0|T1693994400\n" + "button_click@none:1|c|#browser:Chrome,browser.version:1.0|T1693994400\n" + "invalid-metric-name@none:1|c|#browser:Chrome,browser.version:1.0|T1693994400\n" )
476f26ff015e199fef3b9392eb1cd1525d10dbd2
2025-03-20 01:05:21
Dan Fuller
feat(issue-platform): Support passing fingerprints with multiple hashes when sending occurrences (#87311)
false
Support passing fingerprints with multiple hashes when sending occurrences (#87311)
feat
diff --git a/src/sentry/event_manager.py b/src/sentry/event_manager.py index 9fd16a022c55e9..fb236fe9a82e90 100644 --- a/src/sentry/event_manager.py +++ b/src/sentry/event_manager.py @@ -2496,7 +2496,7 @@ def save_grouphash_and_group( event: Event, new_grouphash: str, **group_kwargs: Any, -) -> tuple[Group, bool]: +) -> tuple[Group, bool, GroupHash]: group = None with transaction.atomic(router.db_for_write(GroupHash)): group_hash, created = GroupHash.objects.get_or_create(project=project, hash=new_grouphash) @@ -2510,7 +2510,7 @@ def save_grouphash_and_group( # Group, we can guarantee that the Group will exist at this point and # fetch it via GroupHash group = Group.objects.get(grouphash__project=project, grouphash__hash=new_grouphash) - return group, created + return group, created, group_hash @sentry_sdk.tracing.trace diff --git a/src/sentry/issues/ingest.py b/src/sentry/issues/ingest.py index 26f5513489d0c6..8032ac1ef74b0f 100644 --- a/src/sentry/issues/ingest.py +++ b/src/sentry/issues/ingest.py @@ -74,7 +74,11 @@ def process_occurrence_data(data: dict[str, Any]) -> None: return # Hash fingerprints to make sure they're a consistent length - data["fingerprint"] = [md5(part.encode("utf-8")).hexdigest() for part in data["fingerprint"]] + data["fingerprint"] = hash_fingerprint(data["fingerprint"]) + + +def hash_fingerprint(fingerprint: list[str]) -> list[str]: + return [md5(part.encode("utf-8")).hexdigest() for part in fingerprint] class IssueArgs(TypedDict): @@ -172,20 +176,24 @@ def save_issue_from_occurrence( # until after we have created a `Group`. issue_kwargs["message"] = augment_message_with_occurrence(issue_kwargs["message"], occurrence) - # TODO: For now we will assume a single fingerprint. We can expand later if necessary. - # Note that additional fingerprints won't be used to generated additional issues, they'll be - # used to map the occurrence to a specific issue. - new_grouphash = occurrence.fingerprint[0] - existing_grouphash = ( - GroupHash.objects.filter(project=project, hash=new_grouphash) - .select_related("group") - .first() - ) + existing_grouphashes = { + gh.hash: gh + for gh in GroupHash.objects.filter( + project=project, hash__in=occurrence.fingerprint + ).select_related("group") + } + primary_grouphash = None + for fingerprint_hash in occurrence.fingerprint: + if fingerprint_hash in existing_grouphashes: + primary_grouphash = existing_grouphashes[fingerprint_hash] + break + + if not primary_grouphash: + primary_hash = occurrence.fingerprint[0] - if not existing_grouphash: cluster_key = settings.SENTRY_ISSUE_PLATFORM_RATE_LIMITER_OPTIONS.get("cluster", "default") client = redis.redis_clusters.get(cluster_key) - if not should_create_group(occurrence.type, client, new_grouphash, project): + if not should_create_group(occurrence.type, client, primary_hash, project): metrics.incr("issues.issue.dropped.noise_reduction") return None @@ -213,7 +221,9 @@ def save_issue_from_occurrence( ) as metric_tags, transaction.atomic(router.db_for_write(GroupHash)), ): - group, is_new = save_grouphash_and_group(project, event, new_grouphash, **issue_kwargs) + group, is_new, primary_grouphash = save_grouphash_and_group( + project, event, primary_hash, **issue_kwargs + ) is_regression = False span.set_tag("save_issue_from_occurrence.outcome", "new_group") metric_tags["save_issue_from_occurrence.outcome"] = "new_group" @@ -248,10 +258,10 @@ def save_issue_from_occurrence( except Exception: logger.exception("Failed process assignment for occurrence") - elif existing_grouphash.group is None: + elif primary_grouphash.group is None: return None else: - group = existing_grouphash.group + group = primary_grouphash.group if group.issue_category.value != occurrence.type.category: logger.error( "save_issue_from_occurrence.category_mismatch", @@ -268,6 +278,23 @@ def save_issue_from_occurrence( is_regression = _process_existing_aggregate(group, group_event, issue_kwargs, release) group_info = GroupInfo(group=group, is_new=False, is_regression=is_regression) + additional_hashes = [f for f in occurrence.fingerprint if f != primary_grouphash.hash] + for fingerprint_hash in additional_hashes: + # Attempt to create the additional grouphash links. They shouldn't be linked to other groups, but guard against + # that + group_hash, created = GroupHash.objects.get_or_create( + project=project, hash=fingerprint_hash, defaults={"group": group_info.group} + ) + if not created: + logger.warning( + "Failed to create additional grouphash for group, grouphash associated with existing group", + extra={ + "new_group_id": group_info.group.id, + "hash": fingerprint_hash, + "existing_group_id": group_hash.group_id, + }, + ) + return group_info diff --git a/tests/sentry/event_manager/test_event_manager.py b/tests/sentry/event_manager/test_event_manager.py index 7f3929c60d70ce..5150cf82af3bfe 100644 --- a/tests/sentry/event_manager/test_event_manager.py +++ b/tests/sentry/event_manager/test_event_manager.py @@ -3423,13 +3423,13 @@ def test(self) -> None: perf_data = load_data("transaction-n-plus-one", timestamp=before_now(minutes=10)) event = _get_event_instance(perf_data, project_id=self.project.id) group_hash = "some_group" - group, created = save_grouphash_and_group(self.project, event, group_hash) + group, created, _ = save_grouphash_and_group(self.project, event, group_hash) assert created - group_2, created = save_grouphash_and_group(self.project, event, group_hash) + group_2, created, _ = save_grouphash_and_group(self.project, event, group_hash) assert group.id == group_2.id assert not created assert Group.objects.filter(grouphash__hash=group_hash).count() == 1 - group_3, created = save_grouphash_and_group(self.project, event, "new_hash") + group_3, created, _ = save_grouphash_and_group(self.project, event, "new_hash") assert created assert group_2.id != group_3.id assert Group.objects.filter(grouphash__hash=group_hash).count() == 1 diff --git a/tests/sentry/issues/test_ingest.py b/tests/sentry/issues/test_ingest.py index 90ddde7c94946a..fed907a27c8862 100644 --- a/tests/sentry/issues/test_ingest.py +++ b/tests/sentry/issues/test_ingest.py @@ -19,6 +19,7 @@ ) from sentry.issues.ingest import ( _create_issue_kwargs, + hash_fingerprint, materialize_metadata, save_issue_from_occurrence, save_issue_occurrence, @@ -28,6 +29,7 @@ from sentry.models.group import Group from sentry.models.groupassignee import GroupAssignee from sentry.models.groupenvironment import GroupEnvironment +from sentry.models.grouphash import GroupHash from sentry.models.grouprelease import GroupRelease from sentry.models.release import Release from sentry.models.releaseprojectenvironment import ReleaseProjectEnvironment @@ -213,6 +215,21 @@ def test_new_group(self) -> None: }, ) + def test_new_group_multiple_fingerprint(self) -> None: + fingerprint = ["hi", "bye"] + occurrence = self.build_occurrence(type=ErrorGroupType.type_id, fingerprint=fingerprint) + event = self.store_event(project_id=self.project.id, data={}) + + group_info = save_issue_from_occurrence(occurrence, event, None) + assert group_info is not None + assert group_info.is_new + assert not group_info.is_regression + + group = group_info.group + assert group.title == occurrence.issue_title + grouphashes = set(GroupHash.objects.filter(group=group).values_list("hash", flat=True)) + assert set(hash_fingerprint(fingerprint)) == grouphashes + def test_existing_group(self) -> None: event = self.store_event(data={}, project_id=self.project.id) occurrence = self.build_occurrence(fingerprint=["some-fingerprint"]) @@ -237,6 +254,77 @@ def test_existing_group(self) -> None: assert updated_group.times_seen == 2 assert updated_group.message == "<unlabeled event> new title new subtitle api/123" + def test_existing_group_multiple_fingerprints(self) -> None: + fingerprint = ["some-fingerprint"] + event = self.store_event(data={}, project_id=self.project.id) + occurrence = self.build_occurrence(fingerprint=fingerprint) + group_info = save_issue_from_occurrence(occurrence, event, None) + assert group_info is not None + assert group_info.is_new + grouphashes = set( + GroupHash.objects.filter(group=group_info.group).values_list("hash", flat=True) + ) + assert set(hash_fingerprint(fingerprint)) == grouphashes + + fingerprint = ["some-fingerprint", "another-fingerprint"] + new_event = self.store_event(data={}, project_id=self.project.id) + new_occurrence = self.build_occurrence(fingerprint=fingerprint) + with self.tasks(): + updated_group_info = save_issue_from_occurrence(new_occurrence, new_event, None) + assert updated_group_info is not None + assert group_info.group.id == updated_group_info.group.id + assert not updated_group_info.is_new + assert not updated_group_info.is_regression + grouphashes = set( + GroupHash.objects.filter(group=group_info.group).values_list("hash", flat=True) + ) + assert set(hash_fingerprint(fingerprint)) == grouphashes + + def test_existing_group_multiple_fingerprints_overlap(self) -> None: + fingerprint = ["some-fingerprint"] + group_info = save_issue_from_occurrence( + self.build_occurrence(fingerprint=fingerprint), + self.store_event(data={}, project_id=self.project.id), + None, + ) + assert group_info is not None + assert group_info.is_new + grouphashes = set( + GroupHash.objects.filter(group=group_info.group).values_list("hash", flat=True) + ) + assert set(hash_fingerprint(fingerprint)) == grouphashes + other_fingerprint = ["another-fingerprint"] + other_group_info = save_issue_from_occurrence( + self.build_occurrence(fingerprint=other_fingerprint), + self.store_event(data={}, project_id=self.project.id), + None, + ) + assert other_group_info is not None + assert other_group_info.is_new + grouphashes = set( + GroupHash.objects.filter(group=other_group_info.group).values_list("hash", flat=True) + ) + assert set(hash_fingerprint(other_fingerprint)) == grouphashes + + # Should process the in order, and not join an already used fingerprint + overlapping_fingerprint = ["another-fingerprint", "some-fingerprint"] + new_event = self.store_event(data={}, project_id=self.project.id) + new_occurrence = self.build_occurrence(fingerprint=overlapping_fingerprint) + with self.tasks(): + overlapping_group_info = save_issue_from_occurrence(new_occurrence, new_event, None) + assert overlapping_group_info is not None + assert other_group_info.group.id == overlapping_group_info.group.id + assert not overlapping_group_info.is_new + assert not overlapping_group_info.is_regression + grouphashes = set( + GroupHash.objects.filter(group=group_info.group).values_list("hash", flat=True) + ) + assert set(hash_fingerprint(fingerprint)) == grouphashes + other_grouphashes = set( + GroupHash.objects.filter(group=other_group_info.group).values_list("hash", flat=True) + ) + assert set(hash_fingerprint(other_fingerprint)) == other_grouphashes + def test_existing_group_different_category(self) -> None: event = self.store_event(data={}, project_id=self.project.id) occurrence = self.build_occurrence(fingerprint=["some-fingerprint"])
c81e326cc6c2999cc937f13b6d051adb9d146efa
2023-10-31 23:51:33
anthony sottile
ref: match BaseManager type signatures (#59100)
false
match BaseManager type signatures (#59100)
ref
diff --git a/pyproject.toml b/pyproject.toml index ca4c41424ff5a4..4d9dc35131669a 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -190,7 +190,6 @@ module = [ "sentry.api.endpoints.notifications.notification_actions_available", "sentry.api.endpoints.notifications.notification_actions_details", "sentry.api.endpoints.organization_access_request_details", - "sentry.api.endpoints.organization_auth_provider_send_reminders", "sentry.api.endpoints.organization_code_mapping_codeowners", "sentry.api.endpoints.organization_code_mapping_details", "sentry.api.endpoints.organization_code_mappings", diff --git a/src/sentry/db/models/manager/base.py b/src/sentry/db/models/manager/base.py index 02fbbcb7742270..27f8cf78b8087d 100644 --- a/src/sentry/db/models/manager/base.py +++ b/src/sentry/db/models/manager/base.py @@ -252,13 +252,10 @@ def __value_for_field(self, instance: M, key: str) -> Any: field = instance._meta.get_field(key) return getattr(instance, field.attname) - def contribute_to_class(self, model: M, name: str) -> None: + def contribute_to_class(self, model: type[Model], name: str) -> None: super().contribute_to_class(model, name) class_prepared.connect(self.__class_prepared, sender=model) - def get(self, *args: Any, **kwargs: Any) -> M: - return super().get(*args, **kwargs) - @django_test_transaction_water_mark() def get_from_cache( self, use_replica: bool = settings.SENTRY_MODEL_CACHE_USE_REPLICA, **kwargs: Any diff --git a/src/sentry/db/models/manager/option.py b/src/sentry/db/models/manager/option.py index a60fc136522329..cc63fcc12034ab 100644 --- a/src/sentry/db/models/manager/option.py +++ b/src/sentry/db/models/manager/option.py @@ -1,7 +1,10 @@ +from __future__ import annotations + from typing import Any, Dict, Union from celery.signals import task_postrun from django.core.signals import request_finished +from django.db.models import Model from sentry.db.models.manager import M from sentry.db.models.manager.base import BaseManager, _local_cache @@ -18,7 +21,7 @@ def _option_cache(self) -> Dict[str, Dict[str, Any]]: def clear_local_cache(self, **kwargs: Any) -> None: self._option_cache.clear() - def contribute_to_class(self, model: M, name: str) -> None: + def contribute_to_class(self, model: type[Model], name: str) -> None: super().contribute_to_class(model, name) task_postrun.connect(self.clear_local_cache) request_finished.connect(self.clear_local_cache) diff --git a/src/sentry/models/files/abstractfileblob.py b/src/sentry/models/files/abstractfileblob.py index bf37a9e00c5f1a..0ffb4db4b8569c 100644 --- a/src/sentry/models/files/abstractfileblob.py +++ b/src/sentry/models/files/abstractfileblob.py @@ -206,7 +206,7 @@ def from_file(cls, fileobj, logger=nooplogger) -> Self: # see `_save_blob` above metrics.incr("filestore.upload_race", sample_rate=1.0) saved_path = blob.path - blob = cls.objects.get(checksum=checksum) # type:ignore + blob = cls.objects.get(checksum=checksum) storage.delete(saved_path) metrics.timing("filestore.blob-size", size)
9af2e25b93239946d030acc4041575e28c14e27f
2023-10-02 18:48:27
ArthurKnaus
feat(onboarding): Simplify multi snippet docs javascript [part 1] (#57199)
false
Simplify multi snippet docs javascript [part 1] (#57199)
feat
diff --git a/static/app/gettingStartedDocs/electron/electron.tsx b/static/app/gettingStartedDocs/electron/electron.tsx index d14253e8fc44d5..206c74147f0f74 100644 --- a/static/app/gettingStartedDocs/electron/electron.tsx +++ b/static/app/gettingStartedDocs/electron/electron.tsx @@ -23,14 +23,20 @@ export const steps = ({ description: t('Add the Sentry Electron SDK package as a dependency:'), configurations: [ { - language: 'bash', - code: ` -# Using yarn -yarn add @sentry/electron - -# Using npm -npm install --save @sentry/electron - `, + code: [ + { + label: 'npm', + value: 'npm', + language: 'bash', + code: 'npm install --save @sentry/electron', + }, + { + label: 'yarn', + value: 'yarn', + language: 'bash', + code: 'yarn add @sentry/electron', + }, + ], }, ], }, diff --git a/static/app/gettingStartedDocs/javascript/angular.spec.tsx b/static/app/gettingStartedDocs/javascript/angular.spec.tsx index 6f68aaca9bd900..5ce43e163faf71 100644 --- a/static/app/gettingStartedDocs/javascript/angular.spec.tsx +++ b/static/app/gettingStartedDocs/javascript/angular.spec.tsx @@ -3,7 +3,7 @@ import {render, screen} from 'sentry-test/reactTestingLibrary'; import {StepTitle} from 'sentry/components/onboarding/gettingStartedDoc/step'; import {ProductSolution} from 'sentry/components/onboarding/productSelection'; -import {GettingStartedWithAngular, nextSteps, steps} from './angular'; +import {AngularVersion, GettingStartedWithAngular, nextSteps, steps} from './angular'; describe('GettingStartedWithAngular', function () { it('all products are selected', function () { @@ -19,7 +19,11 @@ describe('GettingStartedWithAngular', function () { ); // Steps - for (const step of steps()) { + for (const step of steps({ + angularVersion: AngularVersion.V12, + errorHandlerProviders: 'test-error-handler-providers', + sentryInitContent: 'test-init-content', + })) { expect( screen.getByRole('heading', {name: step.title ?? StepTitle[step.type]}) ).toBeInTheDocument(); diff --git a/static/app/gettingStartedDocs/javascript/angular.tsx b/static/app/gettingStartedDocs/javascript/angular.tsx index 8eb99a44965172..deb466cb42f54e 100644 --- a/static/app/gettingStartedDocs/javascript/angular.tsx +++ b/static/app/gettingStartedDocs/javascript/angular.tsx @@ -1,26 +1,49 @@ -import styled from '@emotion/styled'; - -import List from 'sentry/components/list/'; -import ListItem from 'sentry/components/list/listItem'; import {Layout, LayoutProps} from 'sentry/components/onboarding/gettingStartedDoc/layout'; import {ModuleProps} from 'sentry/components/onboarding/gettingStartedDoc/sdkDocumentation'; import {StepType} from 'sentry/components/onboarding/gettingStartedDoc/step'; import {getUploadSourceMapsStep} from 'sentry/components/onboarding/gettingStartedDoc/utils'; +import { + PlatformOption, + useUrlPlatformOptions, +} from 'sentry/components/onboarding/platformOptionsControl'; import {ProductSolution} from 'sentry/components/onboarding/productSelection'; import {t, tct} from 'sentry/locale'; -import {space} from 'sentry/styles/space'; import type {Organization, PlatformKey} from 'sentry/types'; +export enum AngularVersion { + V10 = 'v10', + V12 = 'v12', +} + +type PlaformOptionKey = 'angularVersion'; + type StepProps = { + angularVersion: AngularVersion; errorHandlerProviders: string; - newOrg: boolean; - organization: Organization; - platformKey: PlatformKey; - projectId: string; sentryInitContent: string; + newOrg?: boolean; + organization?: Organization; + platformKey?: PlatformKey; + projectId?: string; }; // Configuration Start +const platformOptions: Record<PlaformOptionKey, PlatformOption> = { + angularVersion: { + label: t('Spring Boot Version'), + items: [ + { + label: t('Angular 12+'), + value: AngularVersion.V12, + }, + { + label: t('Angular 10 and 11'), + value: AngularVersion.V10, + }, + ], + }, +}; + const replayIntegration = ` new Sentry.Replay(), `; @@ -57,57 +80,48 @@ const performanceErrorHandler = ` }, `; +function getNpmPackage(angularVersion: AngularVersion) { + return angularVersion === AngularVersion.V12 + ? '@sentry/angular-ivy' + : '@sentry/angular'; +} + export const steps = ({ sentryInitContent, errorHandlerProviders, + angularVersion, ...props -}: Partial<StepProps> = {}): LayoutProps['steps'] => [ +}: StepProps): LayoutProps['steps'] => [ { type: StepType.INSTALL, description: ( - <InstallDescription> - <p> - {tct( - "To use Sentry with your Angular application, you'll need [sentryAngularIvyCode:@sentry/angular-ivy] or [sentryAngularCode:@sentry/angular], Sentry’s Browser Angular SDKs:", - { - sentryAngularIvyCode: <code />, - sentryAngularCode: <code />, - } - )} - </p> - <List symbol="bullet"> - <ListItem> - {tct("If you're using Angular 12 or newer, use [code:@sentry/angular-ivy]", { - code: <code />, - })} - </ListItem> - <ListItem> - {tct("If you're using Angular 10 or 11, use [code:@sentry/angular]", { - code: <code />, - })} - </ListItem> - </List> - <p> - {tct('Add the Sentry SDK as a dependency using [code:yarn] or [code:npm]:', { - code: <code />, - })} - </p> - </InstallDescription> + <p> + {tct( + 'Add the Sentry SDK as a dependency using [codeNpm:npm] or [codeYarn:yarn]:', + { + codeYarn: <code />, + codeNpm: <code />, + } + )} + </p> ), configurations: [ { language: 'bash', - code: ` -# Using yarn (Angular 12+) -yarn add @sentry/angular-ivy -# Using yarn (Angular 10 and 11) -yarn add @sentry/angular - -# Using npm (Angular 12+) -npm install --save @sentry/angular-ivy -# Using npm (Angular 10 and 11) -npm install --save @sentry/angular - `, + code: [ + { + label: 'npm', + value: 'npm', + language: 'bash', + code: `npm install --save ${getNpmPackage(angularVersion)}`, + }, + { + label: 'yarn', + value: 'yarn', + language: 'bash', + code: `yarn add ${getNpmPackage(angularVersion)}`, + }, + ], }, ], }, @@ -120,23 +134,22 @@ npm install --save @sentry/angular { language: 'javascript', code: ` - import { enableProdMode } from "@angular/core"; - import { platformBrowserDynamic } from "@angular/platform-browser-dynamic"; - // import * as Sentry from "@sentry/angular" // for Angular 10/11 instead - import * as Sentry from "@sentry/angular-ivy"; +import { enableProdMode } from "@angular/core"; +import { platformBrowserDynamic } from "@angular/platform-browser-dynamic"; +import * as Sentry from "${getNpmPackage(angularVersion)}"; - import { AppModule } from "./app/app.module"; +import { AppModule } from "./app/app.module"; - Sentry.init({ - ${sentryInitContent} - }); +Sentry.init({ + ${sentryInitContent} +}); - enableProdMode(); - platformBrowserDynamic() - .bootstrapModule(AppModule) - .then((success) => console.log('Bootstrap success')) - .catch((err) => console.error(err)); - `, +enableProdMode(); +platformBrowserDynamic() + .bootstrapModule(AppModule) + .then((success) => console.log('Bootstrap success')) + .catch((err) => console.error(err)); +`, }, { description: t( @@ -144,25 +157,23 @@ npm install --save @sentry/angular ), language: 'javascript', code: ` - import { APP_INITIALIZER, ErrorHandler, NgModule } from "@angular/core"; - import { Router } from "@angular/router"; - // import * as Sentry from "@sentry/angular" // for Angular 10/11 instead - import * as Sentry from "@sentry/angular-ivy"; +import { APP_INITIALIZER, ErrorHandler, NgModule } from "@angular/core"; +import { Router } from "@angular/router"; +import * as Sentry from "${getNpmPackage(angularVersion)}"; - @NgModule({ - // ... - providers: [ - { - provide: ErrorHandler, - useValue: Sentry.createErrorHandler({ - showDialog: true, - }), - },${errorHandlerProviders} - ], - // ... - }) - export class AppModule {} - `, +@NgModule({ + // ... + providers: [ + { + provide: ErrorHandler, + useValue: Sentry.createErrorHandler({ + showDialog: true, + }), + },${errorHandlerProviders} + ], + // ... +}) +export class AppModule {}`, }, ], }, @@ -219,6 +230,7 @@ export function GettingStartedWithAngular({ projectId, ...props }: ModuleProps) { + const optionValues = useUrlPlatformOptions(platformOptions); const integrations: string[] = []; const otherConfigs: string[] = []; @@ -257,12 +269,14 @@ export function GettingStartedWithAngular({ steps={steps({ sentryInitContent: sentryInitContent.join('\n'), errorHandlerProviders: errorHandlerProviders.join('\n'), + angularVersion: optionValues.angularVersion as AngularVersion, organization, newOrg, platformKey, projectId, })} nextSteps={nextStepDocs} + platformOptions={platformOptions} newOrg={newOrg} platformKey={platformKey} {...props} @@ -271,9 +285,3 @@ export function GettingStartedWithAngular({ } export default GettingStartedWithAngular; - -const InstallDescription = styled('div')` - display: flex; - flex-direction: column; - gap: ${space(1)}; -`; diff --git a/static/app/gettingStartedDocs/javascript/gatsby.tsx b/static/app/gettingStartedDocs/javascript/gatsby.tsx index 35eea877c3cd40..947e7c9a2f10c1 100644 --- a/static/app/gettingStartedDocs/javascript/gatsby.tsx +++ b/static/app/gettingStartedDocs/javascript/gatsby.tsx @@ -43,19 +43,34 @@ export const steps = ({ }: Partial<StepProps> = {}): LayoutProps['steps'] => [ { type: StepType.INSTALL, - description: t( - 'Sentry captures data by using an SDK within your application’s runtime.' + description: ( + <p> + {tct( + 'Add the Sentry SDK as a dependency using [codeNpm:npm] or [codeYarn:yarn]:', + { + codeYarn: <code />, + codeNpm: <code />, + } + )} + </p> ), configurations: [ { language: 'bash', - code: ` -# Using yarn -yarn add @sentry/gatsby - -# Using npm -npm install --save @sentry/gatsby - `, + code: [ + { + label: 'npm', + value: 'npm', + language: 'bash', + code: 'npm install --save @sentry/gatsby', + }, + { + label: 'yarn', + value: 'yarn', + language: 'bash', + code: 'yarn add @sentry/gatsby', + }, + ], }, ], }, diff --git a/static/app/gettingStartedDocs/javascript/javascript.tsx b/static/app/gettingStartedDocs/javascript/javascript.tsx index d1e1f328f60414..5d119a5ea4a548 100644 --- a/static/app/gettingStartedDocs/javascript/javascript.tsx +++ b/static/app/gettingStartedDocs/javascript/javascript.tsx @@ -49,13 +49,20 @@ export const steps = ({ configurations: [ { language: 'bash', - code: ` -# Using yarn -yarn add @sentry/browser - -# Using npm -npm install --save @sentry/browser - `, + code: [ + { + label: 'npm', + value: 'npm', + language: 'bash', + code: 'npm install --save @sentry/electron', + }, + { + label: 'yarn', + value: 'yarn', + language: 'bash', + code: 'yarn add @sentry/electron', + }, + ], }, ], }, diff --git a/static/app/gettingStartedDocs/javascript/react.tsx b/static/app/gettingStartedDocs/javascript/react.tsx index 5ff88d94eff23b..83b17e6e8e38bf 100644 --- a/static/app/gettingStartedDocs/javascript/react.tsx +++ b/static/app/gettingStartedDocs/javascript/react.tsx @@ -3,7 +3,7 @@ import {ModuleProps} from 'sentry/components/onboarding/gettingStartedDoc/sdkDoc import {StepType} from 'sentry/components/onboarding/gettingStartedDoc/step'; import {getUploadSourceMapsStep} from 'sentry/components/onboarding/gettingStartedDoc/utils'; import {ProductSolution} from 'sentry/components/onboarding/productSelection'; -import {t} from 'sentry/locale'; +import {t, tct} from 'sentry/locale'; import type {Organization, PlatformKey} from 'sentry/types'; type StepProps = { @@ -43,19 +43,34 @@ export const steps = ({ }: Partial<StepProps> = {}): LayoutProps['steps'] => [ { type: StepType.INSTALL, - description: t( - 'Sentry captures data by using an SDK within your application’s runtime.' + description: ( + <p> + {tct( + 'Add the Sentry SDK as a dependency using [codeNpm:npm] or [codeYarn:yarn]:', + { + codeYarn: <code />, + codeNpm: <code />, + } + )} + </p> ), configurations: [ { language: 'bash', - code: ` -# Using yarn -yarn add @sentry/react - -# Using npm -npm install --save @sentry/react - `, + code: [ + { + label: 'npm', + value: 'npm', + language: 'bash', + code: 'npm install --save @sentry/react', + }, + { + label: 'yarn', + value: 'yarn', + language: 'bash', + code: 'yarn add @sentry/react', + }, + ], }, ], }, diff --git a/static/app/gettingStartedDocs/javascript/remix.tsx b/static/app/gettingStartedDocs/javascript/remix.tsx index 455f97f465d44f..84e0756876ac6c 100644 --- a/static/app/gettingStartedDocs/javascript/remix.tsx +++ b/static/app/gettingStartedDocs/javascript/remix.tsx @@ -46,19 +46,34 @@ export const steps = ({ } = {}): LayoutProps['steps'] => [ { type: StepType.INSTALL, - description: t( - 'Sentry captures data by using an SDK within your application’s runtime.' + description: ( + <p> + {tct( + 'Add the Sentry SDK as a dependency using [codeNpm:npm] or [codeYarn:yarn]:', + { + codeYarn: <code />, + codeNpm: <code />, + } + )} + </p> ), configurations: [ { language: 'bash', - code: ` -# Using yarn -yarn add @sentry/remix - -# Using npm -npm install --save @sentry/remix - `, + code: [ + { + label: 'npm', + value: 'npm', + language: 'bash', + code: 'npm install --save @sentry/remix', + }, + { + label: 'yarn', + value: 'yarn', + language: 'bash', + code: 'yarn add @sentry/remix', + }, + ], }, ], }, diff --git a/static/app/gettingStartedDocs/javascript/svelte.tsx b/static/app/gettingStartedDocs/javascript/svelte.tsx index 246abc5c9fc64f..751e6042f6b6ed 100644 --- a/static/app/gettingStartedDocs/javascript/svelte.tsx +++ b/static/app/gettingStartedDocs/javascript/svelte.tsx @@ -43,19 +43,34 @@ export const steps = ({ }: Partial<StepProps> = {}): LayoutProps['steps'] => [ { type: StepType.INSTALL, - description: t( - 'Sentry captures data by using an SDK within your application’s runtime.' + description: ( + <p> + {tct( + 'Add the Sentry SDK as a dependency using [codeNpm:npm] or [codeYarn:yarn]:', + { + codeYarn: <code />, + codeNpm: <code />, + } + )} + </p> ), configurations: [ { language: 'bash', - code: ` -# Using yarn -yarn add @sentry/svelte - -# Using npm -npm install --save @sentry/svelte - `, + code: [ + { + label: 'npm', + value: 'npm', + language: 'bash', + code: 'npm install --save @sentry/svelte', + }, + { + label: 'yarn', + value: 'yarn', + language: 'bash', + code: 'yarn add @sentry/svelte', + }, + ], }, ], }, diff --git a/static/app/gettingStartedDocs/javascript/vue.spec.tsx b/static/app/gettingStartedDocs/javascript/vue.spec.tsx index 7b3ea74e0dac00..e63fd46ab65f44 100644 --- a/static/app/gettingStartedDocs/javascript/vue.spec.tsx +++ b/static/app/gettingStartedDocs/javascript/vue.spec.tsx @@ -3,7 +3,7 @@ import {render, screen} from 'sentry-test/reactTestingLibrary'; import {StepTitle} from 'sentry/components/onboarding/gettingStartedDoc/step'; import {ProductSolution} from 'sentry/components/onboarding/productSelection'; -import {GettingStartedWithVue, nextSteps, steps} from './vue'; +import {GettingStartedWithVue, nextSteps, steps, VueVersion} from './vue'; describe('GettingStartedWithVue', function () { it('all products are selected', function () { @@ -19,7 +19,10 @@ describe('GettingStartedWithVue', function () { ); // Steps - for (const step of steps()) { + for (const step of steps({ + vueVersion: VueVersion.V3, + sentryInitContent: 'test-init-content', + })) { expect( screen.getByRole('heading', {name: step.title ?? StepTitle[step.type]}) ).toBeInTheDocument(); diff --git a/static/app/gettingStartedDocs/javascript/vue.tsx b/static/app/gettingStartedDocs/javascript/vue.tsx index 284237910d3e12..1115f409829283 100644 --- a/static/app/gettingStartedDocs/javascript/vue.tsx +++ b/static/app/gettingStartedDocs/javascript/vue.tsx @@ -2,19 +2,47 @@ import {Layout, LayoutProps} from 'sentry/components/onboarding/gettingStartedDo import {ModuleProps} from 'sentry/components/onboarding/gettingStartedDoc/sdkDocumentation'; import {StepType} from 'sentry/components/onboarding/gettingStartedDoc/step'; import {getUploadSourceMapsStep} from 'sentry/components/onboarding/gettingStartedDoc/utils'; +import { + PlatformOption, + useUrlPlatformOptions, +} from 'sentry/components/onboarding/platformOptionsControl'; import {ProductSolution} from 'sentry/components/onboarding/productSelection'; -import {t} from 'sentry/locale'; +import {t, tct} from 'sentry/locale'; import type {Organization, PlatformKey} from 'sentry/types'; +export enum VueVersion { + V3 = 'v3', + V2 = 'v2', +} + +type PlaformOptionKey = 'vueVersion'; + type StepProps = { - newOrg: boolean; - organization: Organization; - platformKey: PlatformKey; - projectId: string; sentryInitContent: string; + vueVersion: VueVersion; + newOrg?: boolean; + organization?: Organization; + platformKey?: PlatformKey; + projectId?: string; }; // Configuration Start +const platformOptions: Record<PlaformOptionKey, PlatformOption> = { + vueVersion: { + label: t('Spring Boot Version'), + items: [ + { + label: t('Vue 3'), + value: VueVersion.V3, + }, + { + label: t('Vue 2'), + value: VueVersion.V2, + }, + ], + }, +}; + const replayIntegration = ` new Sentry.Replay(), `; @@ -40,23 +68,39 @@ tracesSampleRate: 1.0, // Capture 100% of the transactions, reduce in production export const steps = ({ sentryInitContent, + vueVersion, ...props -}: Partial<StepProps> = {}): LayoutProps['steps'] => [ +}: StepProps): LayoutProps['steps'] => [ { type: StepType.INSTALL, - description: t( - 'Sentry captures data by using an SDK within your application’s runtime.' + description: ( + <p> + {tct( + 'Add the Sentry SDK as a dependency using [codeNpm:npm] or [codeYarn:yarn]:', + { + codeYarn: <code />, + codeNpm: <code />, + } + )} + </p> ), configurations: [ { language: 'bash', - code: ` -# Using yarn -yarn add @sentry/vue - -# Using npm -npm install --save @sentry/vue - `, + code: [ + { + label: 'npm', + value: 'npm', + language: 'bash', + code: 'npm install --save @sentry/vue', + }, + { + label: 'yarn', + value: 'yarn', + language: 'bash', + code: 'yarn add @sentry/vue', + }, + ], }, ], }, @@ -65,11 +109,12 @@ npm install --save @sentry/vue description: t( "Initialize Sentry as early as possible in your application's lifecycle." ), - configurations: [ - { - description: <h5>Vue 3</h5>, - language: 'javascript', - code: ` + configurations: + vueVersion === VueVersion.V3 + ? [ + { + language: 'javascript', + code: ` import { createApp } from "vue"; import { createRouter } from "vue-router"; import * as Sentry from "@sentry/vue"; @@ -89,11 +134,12 @@ npm install --save @sentry/vue app.use(router); app.mount("#app"); `, - }, - { - description: <h5>Vue 2</h5>, - language: 'javascript', - code: ` + }, + ] + : [ + { + language: 'javascript', + code: ` import Vue from "vue"; import Router from "vue-router"; import * as Sentry from "@sentry/vue"; @@ -116,8 +162,8 @@ npm install --save @sentry/vue render: (h) => h(App), }).$mount("#app"); `, - }, - ], + }, + ], }, getUploadSourceMapsStep({ guideLink: 'https://docs.sentry.io/platforms/javascript/guides/vue/sourcemaps/', @@ -178,6 +224,7 @@ export function GettingStartedWithVue({ projectId, ...props }: ModuleProps) { + const optionValues = useUrlPlatformOptions(platformOptions); const integrations: string[] = []; const otherConfigs: string[] = []; @@ -213,6 +260,7 @@ export function GettingStartedWithVue({ <Layout steps={steps({ sentryInitContent: sentryInitContent.join('\n'), + vueVersion: optionValues.vueVersion as VueVersion, organization, newOrg, platformKey, @@ -221,6 +269,7 @@ export function GettingStartedWithVue({ nextSteps={nextStepDocs} newOrg={newOrg} platformKey={platformKey} + platformOptions={platformOptions} {...props} /> );
5fc860388e0f75cc773c6c51402f8f93a6a42d20
2020-02-20 04:44:43
Billy Vong
fix(workflow): Change "View in Discover" link to include `interval` (#17118)
false
Change "View in Discover" link to include `interval` (#17118)
fix
diff --git a/src/sentry/static/sentry/app/views/alerts/details/body.tsx b/src/sentry/static/sentry/app/views/alerts/details/body.tsx index 54e2fd1e380bbf..f28481921b9342 100644 --- a/src/sentry/static/sentry/app/views/alerts/details/body.tsx +++ b/src/sentry/static/sentry/app/views/alerts/details/body.tsx @@ -1,5 +1,6 @@ import {RouteComponentProps} from 'react-router/lib/Router'; import React from 'react'; +import moment from 'moment-timezone'; import styled from '@emotion/styled'; import { @@ -11,6 +12,7 @@ import {NewQuery, Project} from 'app/types'; import {PageContent} from 'app/styles/organization'; import {defined} from 'app/utils'; import {getDisplayForAlertRuleAggregation} from 'app/views/alerts/utils'; +import {getUtcDateString, intervalToMilliseconds} from 'app/utils/dates'; import {t} from 'app/locale'; import Duration from 'app/components/duration'; import EventView from 'app/views/eventsV2/eventView'; @@ -43,6 +45,19 @@ export default class DetailsBody extends React.Component<Props> { return ''; } + const timeWindowString = `${incident.alertRule.timeWindow}m`; + const timeWindowInMs = intervalToMilliseconds(timeWindowString); + const startBeforeTimeWindow = moment(incident.dateStarted).subtract( + timeWindowInMs, + 'ms' + ); + const end = incident.dateClosed ?? getUtcDateString(new Date()); + + // We want the discover chart to start at "dateStarted" - "timeWindow" - "20%" + const additionalWindowBeforeStart = + moment(end).diff(startBeforeTimeWindow, 'ms') * 0.2; + const start = startBeforeTimeWindow.subtract(additionalWindowBeforeStart, 'ms'); + const discoverQuery: NewQuery = { id: undefined, name: (incident && incident.title) || '', @@ -57,11 +72,17 @@ export default class DetailsBody extends React.Component<Props> { .filter(({slug}) => incident.projects.includes(slug)) .map(({id}) => Number(id)), version: 2 as const, - range: `${incident.alertRule.timeWindow}m`, + start: getUtcDateString(start), + end, }; const discoverView = EventView.fromSavedQuery(discoverQuery); - return discoverView.getResultsViewUrlTarget(orgId); + const {query, ...toObject} = discoverView.getResultsViewUrlTarget(orgId); + + return { + query: {...query, interval: timeWindowString}, + ...toObject, + }; } /**
a67c6711605fc5e04142c7ddc20a7f92c2ed3df9
2018-08-29 18:39:58
Jan Michael Auer
fix(native): Update test fixtures after symbolic update (#9550)
false
Update test fixtures after symbolic update (#9550)
fix
diff --git a/requirements-base.txt b/requirements-base.txt index 20e7f87f63d121..73d6561b8c5cb9 100644 --- a/requirements-base.txt +++ b/requirements-base.txt @@ -63,7 +63,7 @@ sqlparse>=0.1.16,<0.2.0 statsd>=3.1.0,<3.2.0 strict-rfc3339>=0.7 structlog==16.1.0 -symbolic>=5.1.0,<6.0.0 +symbolic>=5.1.1,<6.0.0 toronado>=0.0.11,<0.1.0 ua-parser>=0.6.1,<0.8.0 # for bitbucket client diff --git a/tests/sentry/lang/native/test_utils.py b/tests/sentry/lang/native/test_utils.py index 5ea9843afbb3df..c8d4cac4b543bf 100644 --- a/tests/sentry/lang/native/test_utils.py +++ b/tests/sentry/lang/native/test_utils.py @@ -165,6 +165,20 @@ def test_minidump_linux(): }, 'debug_meta': { 'images': [ + { + 'id': u'c0bcc3f1-9827-fe65-3058-404b2831d9e6', + 'image_addr': '0x400000', + 'image_size': 106496, + 'name': u'/work/linux/build/crash', + 'type': 'symbolic' + }, + { + 'id': u'e45db8df-af2d-09fd-640c-8fe377d572de', + 'image_addr': '0x7f513fe54000', + 'image_size': 1081344, + 'name': u'/lib/x86_64-linux-gnu/libm-2.23.so', + 'type': 'symbolic' + }, { 'id': u'451a38b5-0679-79d2-0738-22a5ceb24c4b', 'image_addr': '0x7f514015d000', @@ -172,6 +186,27 @@ def test_minidump_linux(): 'name': u'/lib/x86_64-linux-gnu/libc-2.23.so', 'type': 'symbolic' }, + { + 'id': u'e20a2268-5dc6-c165-b6aa-a12fa6765a6e', + 'image_addr': '0x7f5140527000', + 'image_size': 90112, + 'name': u'/lib/x86_64-linux-gnu/libgcc_s.so.1', + 'type': 'symbolic' + }, + { + 'id': u'81c893cb-9b92-3c52-01ac-ef171b52d526', + 'image_addr': '0x7f514073d000', + 'image_size': 1515520, + 'name': u'/usr/lib/x86_64-linux-gnu/libstdc++.so.6.0.21', + 'type': 'symbolic' + }, + { + 'id': u'23e017ce-2254-fc65-11d9-bc8f534bb4f0', + 'image_addr': '0x7f5140abf000', + 'image_size': 98304, + 'name': u'/lib/x86_64-linux-gnu/libpthread-2.23.so', + 'type': 'symbolic' + }, { 'id': u'59627b5d-2255-a375-c17b-d4c3fd05f5a6', 'image_addr': '0x7f5140cdc000', @@ -180,10 +215,10 @@ def test_minidump_linux(): 'type': 'symbolic' }, { - 'id': u'c0bcc3f1-9827-fe65-3058-404b2831d9e6', - 'image_addr': '0x400000', - 'image_size': 106496, - 'name': u'/work/linux/build/crash', + 'id': u'75185f6c-04b9-b48f-b8df-d832e74ad31a', + 'image_addr': '0x7fff5aef1000', + 'image_size': 8192, + 'name': u'linux-gate.so', 'type': 'symbolic' } ] @@ -350,12 +385,299 @@ def test_minidump_macos(): 'name': u'/Users/travis/build/getsentry/breakpad-tools/macos/build/./crash', 'type': 'symbolic' }, + { + 'id': u'36385a3a-60d3-32db-bf55-c6d8931a7aa6', + 'image_addr': '0x7fffd229c000', + 'image_size': 4800512, + 'name': u'/System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation', + 'type': 'symbolic' + }, + { + 'id': u'84a04d24-0e60-3810-a8c0-90a65e2df61a', + 'image_addr': '0x7fffe668e000', + 'image_size': 8192, + 'name': u'/usr/lib/libDiagnosticMessagesClient.dylib', + 'type': 'symbolic' + }, + { + 'id': u'f18ac1e7-c6f1-34b1-8069-be571b3231d4', + 'image_addr': '0x7fffe68cd000', + 'image_size': 8192, + 'name': u'/usr/lib/libSystem.B.dylib', + 'type': 'symbolic' + }, + { + 'id': u'0b43bb5d-e6eb-3464-8de9-b41ac8ed9d1c', + 'image_addr': '0x7fffe6a80000', + 'image_size': 356352, + 'name': u'/usr/lib/libc++.1.dylib', + 'type': 'symbolic' + }, + { + 'id': u'bc271ad3-831b-362a-9da7-e8c51f285fe4', + 'image_addr': '0x7fffe6ad7000', + 'image_size': 172032, + 'name': u'/usr/lib/libc++abi.dylib', + 'type': 'symbolic' + }, + { + 'id': u'ccd2ed24-3071-383b-925d-8d763bb12a6f', + 'image_addr': '0x7fffe7041000', + 'image_size': 2252800, + 'name': u'/usr/lib/libicucore.A.dylib', + 'type': 'symbolic' + }, + { + 'id': u'4df3c25c-52c2-3f01-a3ef-0d9d53a73c1c', + 'image_addr': '0x7fffe75f5000', + 'image_size': 4022272, + 'name': u'/usr/lib/libobjc.A.dylib', + 'type': 'symbolic' + }, + { + 'id': u'46e3ffa2-4328-327a-8d34-a03e20bffb8e', + 'image_addr': '0x7fffe7def000', + 'image_size': 73728, + 'name': u'/usr/lib/libz.1.dylib', + 'type': 'symbolic' + }, + { + 'id': u'093a4dab-8385-3d47-a350-e20cb7ccf7bf', + 'image_addr': '0x7fffe7e0f000', + 'image_size': 20480, + 'name': u'/usr/lib/system/libcache.dylib', + 'type': 'symbolic' + }, + { + 'id': u'8a64d1b0-c70e-385c-92f0-e669079fda90', + 'image_addr': '0x7fffe7e14000', + 'image_size': 45056, + 'name': u'/usr/lib/system/libcommonCrypto.dylib', + 'type': 'symbolic' + }, + { + 'id': u'55d47421-772a-32ab-b529-1a46c2f43b4d', + 'image_addr': '0x7fffe7e1f000', + 'image_size': 32768, + 'name': u'/usr/lib/system/libcompiler_rt.dylib', + 'type': 'symbolic' + }, + { + 'id': u'819bea3c-df11-3e3d-a1a1-5a51c5bf1961', + 'image_addr': '0x7fffe7e27000', + 'image_size': 36864, + 'name': u'/usr/lib/system/libcopyfile.dylib', + 'type': 'symbolic' + }, + { + 'id': u'65d7165e-2e71-335d-a2d6-33f78e2df0c1', + 'image_addr': '0x7fffe7e30000', + 'image_size': 540672, + 'name': u'/usr/lib/system/libcorecrypto.dylib', + 'type': 'symbolic' + }, + { + 'id': u'6582bad6-ed27-3b30-b620-90b1c5a4ae3c', + 'image_addr': '0x7fffe7eb4000', + 'image_size': 204800, + 'name': u'/usr/lib/system/libdispatch.dylib', + 'type': 'symbolic' + }, { 'id': u'9b2ac56d-107c-3541-a127-9094a751f2c9', 'image_addr': '0x7fffe7ee6000', 'image_size': 24576, 'name': u'/usr/lib/system/libdyld.dylib', 'type': 'symbolic' + }, + { + 'id': u'7aa011a9-dc21-3488-bf73-3b5b14d1fdd6', + 'image_addr': '0x7fffe7eec000', + 'image_size': 4096, + 'name': u'/usr/lib/system/libkeymgr.dylib', + 'type': 'symbolic' + }, + { + 'id': u'b856abd2-896e-3de0-b2c8-146a6af8e2a7', + 'image_addr': '0x7fffe7efa000', + 'image_size': 4096, + 'name': u'/usr/lib/system/liblaunch.dylib', + 'type': 'symbolic' + }, + { + 'id': u'17d5d855-f6c3-3b04-b680-e9bf02ef8aed', + 'image_addr': '0x7fffe7efb000', + 'image_size': 24576, + 'name': u'/usr/lib/system/libmacho.dylib', + 'type': 'symbolic' + }, + { + 'id': u'12448cc2-378e-35f3-be33-9dc395a5b970', + 'image_addr': '0x7fffe7f01000', + 'image_size': 12288, + 'name': u'/usr/lib/system/libquarantine.dylib', + 'type': 'symbolic' + }, + { + 'id': u'38d4cb9c-10cd-30d3-8b7b-a515ec75fe85', + 'image_addr': '0x7fffe7f04000', + 'image_size': 8192, + 'name': u'/usr/lib/system/libremovefile.dylib', + 'type': 'symbolic' + }, + { + 'id': u'096e4228-3b7c-30a6-8b13-ec909a64499a', + 'image_addr': '0x7fffe7f06000', + 'image_size': 102400, + 'name': u'/usr/lib/system/libsystem_asl.dylib', + 'type': 'symbolic' + }, + { + 'id': u'10dc5404-73ab-35b3-a277-a8afecb476eb', + 'image_addr': '0x7fffe7f1f000', + 'image_size': 4096, + 'name': u'/usr/lib/system/libsystem_blocks.dylib', + 'type': 'symbolic' + }, + { + 'id': u'e5ae5244-7d0c-36ac-8bb6-c7ae7ea52a4b', + 'image_addr': '0x7fffe7f20000', + 'image_size': 581632, + 'name': u'/usr/lib/system/libsystem_c.dylib', + 'type': 'symbolic' + }, + { + 'id': u'becc01a2-ca8d-31e6-bcdf-d452965fa976', + 'image_addr': '0x7fffe7fae000', + 'image_size': 16384, + 'name': u'/usr/lib/system/libsystem_configuration.dylib', + 'type': 'symbolic' + }, + { + 'id': u'7d26de79-b424-3450-85e1-f7fab32714ab', + 'image_addr': '0x7fffe7fb2000', + 'image_size': 16384, + 'name': u'/usr/lib/system/libsystem_coreservices.dylib', + 'type': 'symbolic' + }, + { + 'id': u'ec6fcf07-dcfb-3a03-9cc9-6dd3709974c6', + 'image_addr': '0x7fffe7fb6000', + 'image_size': 102400, + 'name': u'/usr/lib/system/libsystem_coretls.dylib', + 'type': 'symbolic' + }, + { + 'id': u'cc960215-0b1b-3822-a13a-3dde96fa796f', + 'image_addr': '0x7fffe7fcf000', + 'image_size': 28672, + 'name': u'/usr/lib/system/libsystem_dnssd.dylib', + 'type': 'symbolic' + }, + { + 'id': u'611db84c-bf70-3f92-8702-b9f28a900920', + 'image_addr': '0x7fffe7fd6000', + 'image_size': 172032, + 'name': u'/usr/lib/system/libsystem_info.dylib', + 'type': 'symbolic' + }, + { + 'id': u'34b1f16c-bc9c-3c5f-9045-0cae91cb5914', + 'image_addr': '0x7fffe8000000', + 'image_size': 143360, + 'name': u'/usr/lib/system/libsystem_kernel.dylib', + 'type': 'symbolic' + }, + { + 'id': u'86d499b5-bbdc-3d3b-8a4e-97ae8e6672a4', + 'image_addr': '0x7fffe8023000', + 'image_size': 294912, + 'name': u'/usr/lib/system/libsystem_m.dylib', + 'type': 'symbolic' + }, + { + 'id': u'a3d15f17-99a6-3367-8c7e-4280e8619c95', + 'image_addr': '0x7fffe806b000', + 'image_size': 126976, + 'name': u'/usr/lib/system/libsystem_malloc.dylib', + 'type': 'symbolic' + }, + { + 'id': u'369d0221-56ca-3c3e-9ede-94b41cae77b7', + 'image_addr': '0x7fffe808a000', + 'image_size': 368640, + 'name': u'/usr/lib/system/libsystem_network.dylib', + 'type': 'symbolic' + }, + { + 'id': u'b021f2b3-8a75-3633-abb0-fc012b8e9b0c', + 'image_addr': '0x7fffe80e4000', + 'image_size': 40960, + 'name': u'/usr/lib/system/libsystem_networkextension.dylib', + 'type': 'symbolic' + }, + { + 'id': u'b8160190-a069-3b3a-bdf6-2aa408221fae', + 'image_addr': '0x7fffe80ee000', + 'image_size': 40960, + 'name': u'/usr/lib/system/libsystem_notify.dylib', + 'type': 'symbolic' + }, + { + 'id': u'897462fd-b318-321b-a554-e61982630f7e', + 'image_addr': '0x7fffe80f8000', + 'image_size': 36864, + 'name': u'/usr/lib/system/libsystem_platform.dylib', + 'type': 'symbolic' + }, + { + 'id': u'b8fb5e20-3295-39e2-b5eb-b464d1d4b104', + 'image_addr': '0x7fffe8101000', + 'image_size': 45056, + 'name': u'/usr/lib/system/libsystem_pthread.dylib', + 'type': 'symbolic' + }, + { + 'id': u'4b92ec49-acd0-36ae-b07a-a2b8152eaf9d', + 'image_addr': '0x7fffe810c000', + 'image_size': 16384, + 'name': u'/usr/lib/system/libsystem_sandbox.dylib', + 'type': 'symbolic' + }, + { + 'id': u'f78b847b-3565-3e4b-98a6-f7ad40392e2d', + 'image_addr': '0x7fffe8110000', + 'image_size': 8192, + 'name': u'/usr/lib/system/libsystem_secinit.dylib', + 'type': 'symbolic' + }, + { + 'id': u'3390e07c-c1ce-348f-adbd-2c5440b45eaa', + 'image_addr': '0x7fffe8112000', + 'image_size': 32768, + 'name': u'/usr/lib/system/libsystem_symptoms.dylib', + 'type': 'symbolic' + }, + { + 'id': u'ac63a7fe-50d9-3a30-96e6-f6b7ff16e465', + 'image_addr': '0x7fffe811a000', + 'image_size': 81920, + 'name': u'/usr/lib/system/libsystem_trace.dylib', + 'type': 'symbolic' + }, + { + 'id': u'3d50d8a8-c460-334d-a519-2da841102c6b', + 'image_addr': '0x7fffe812e000', + 'image_size': 24576, + 'name': u'/usr/lib/system/libunwind.dylib', + 'type': 'symbolic' + }, + { + 'id': u'bf896df0-d8e9-31a8-a4b3-01120bfeee52', + 'image_addr': '0x7fffe8134000', + 'image_size': 172032, + 'name': u'/usr/lib/system/libxpc.dylib', + 'type': 'symbolic' } ] }, @@ -452,17 +774,24 @@ def test_minidump_windows(): 'type': 'symbolic' }, { - 'id': u'971f98e5-ce60-41ff-b2d7-235bbeb34578-1', - 'image_addr': '0x77170000', - 'image_size': 1585152, - 'name': u'C:\\Windows\\System32\\ntdll.dll', + 'id': u'9c2a902b-6fdf-40ad-8308-588a41d572a0-1', + 'image_addr': '0x70850000', + 'image_size': 1331200, + 'name': u'C:\\Windows\\System32\\dbghelp.dll', 'type': 'symbolic' }, { - 'id': u'ae131c67-27a7-4fa1-9916-b5a4aef41190-1', - 'image_addr': '0x75810000', - 'image_size': 790528, - 'name': u'C:\\Windows\\System32\\rpcrt4.dll', + 'id': u'bf5257f7-8c26-43dd-9bb7-901625e1136a-1', + 'image_addr': '0x709a0000', + 'image_size': 442368, + 'name': u'C:\\Windows\\System32\\msvcp140.dll', + 'type': 'symbolic' + }, + { + 'id': u'8daf7773-372f-460a-af38-944e193f7e33-1', + 'image_addr': '0x70a10000', + 'image_size': 598016, + 'name': u'C:\\Windows\\System32\\apphelp.dll', 'type': 'symbolic' }, { @@ -472,12 +801,89 @@ def test_minidump_windows(): 'name': u'C:\\Windows\\System32\\dbgcore.dll', 'type': 'symbolic' }, + { + 'id': u'0ed80a50-ecda-472b-86a4-eb6c833f8e1b-1', + 'image_addr': '0x70c60000', + 'image_size': 81920, + 'name': u'C:\\Windows\\System32\\VCRUNTIME140.dll', + 'type': 'symbolic' + }, + { + 'id': u'147c51fb-7ca1-408f-85b5-285f2ad6f9c5-1', + 'image_addr': '0x73ba0000', + 'image_size': 40960, + 'name': u'C:\\Windows\\System32\\CRYPTBASE.dll', + 'type': 'symbolic' + }, + { + 'id': u'51e432b1-0450-4b19-8ed1-6d4335f9f543-1', + 'image_addr': '0x73bb0000', + 'image_size': 126976, + 'name': u'C:\\Windows\\System32\\sspicli.dll', + 'type': 'symbolic' + }, + { + 'id': u'0c799483-b549-417d-8433-4331852031fe-1', + 'image_addr': '0x73c70000', + 'image_size': 487424, + 'name': u'C:\\Windows\\System32\\advapi32.dll', + 'type': 'symbolic' + }, + { + 'id': u'6f6409b3-d520-43c7-9b2f-62e00bfe761c-1', + 'image_addr': '0x73cf0000', + 'image_size': 778240, + 'name': u'C:\\Windows\\System32\\msvcrt.dll', + 'type': 'symbolic' + }, + { + 'id': u'6f6a05dd-0a80-478b-a419-9b88703bf75b-1', + 'image_addr': '0x74450000', + 'image_size': 266240, + 'name': u'C:\\Windows\\System32\\sechost.dll', + 'type': 'symbolic' + }, { 'id': u'd3474559-96f7-47d6-bf43-c176b2171e68-1', 'image_addr': '0x75050000', 'image_size': 917504, 'name': u'C:\\Windows\\System32\\kernel32.dll', 'type': 'symbolic' + }, + { + 'id': u'287b19c3-9209-4a2b-bb8f-bcc37f411b11-1', + 'image_addr': '0x75130000', + 'image_size': 368640, + 'name': u'C:\\Windows\\System32\\bcryptPrimitives.dll', + 'type': 'symbolic' + }, + { + 'id': u'ae131c67-27a7-4fa1-9916-b5a4aef41190-1', + 'image_addr': '0x75810000', + 'image_size': 790528, + 'name': u'C:\\Windows\\System32\\rpcrt4.dll', + 'type': 'symbolic' + }, + { + 'id': u'6bedcbce-0a3a-40e9-8040-81c2c8c6cc2f-1', + 'image_addr': '0x758f0000', + 'image_size': 917504, + 'name': u'C:\\Windows\\System32\\ucrtbase.dll', + 'type': 'symbolic' + }, + { + 'id': u'8462294a-c645-402d-ac82-a4e95f61ddf9-1', + 'image_addr': '0x76db0000', + 'image_size': 1708032, + 'name': u'C:\\Windows\\System32\\KERNELBASE.dll', + 'type': 'symbolic' + }, + { + 'id': u'971f98e5-ce60-41ff-b2d7-235bbeb34578-1', + 'image_addr': '0x77170000', + 'image_size': 1585152, + 'name': u'C:\\Windows\\System32\\ntdll.dll', + 'type': 'symbolic' } ] },
1bf83a66c24bfe56fb1c9c27472845e64dcc9f6b
2022-10-27 13:43:45
Roman Zavarnitsyn
chore(events): Use longKeys for Java packages (#40539)
false
Use longKeys for Java packages (#40539)
chore
diff --git a/static/app/components/events/packageData.tsx b/static/app/components/events/packageData.tsx index b374e31ec5e66f..d4dbd72372b364 100644 --- a/static/app/components/events/packageData.tsx +++ b/static/app/components/events/packageData.tsx @@ -24,6 +24,10 @@ export function EventPackageData({event}: Props) { longKeys = true; title = t('Assemblies'); break; + case 'java': + longKeys = true; + title = t('Dependencies'); + break; default: longKeys = false; title = t('Packages');
47a6501c0150b020b9d515e9171a7b9f4bc66f5e
2020-02-28 06:20:28
Evan Purkhiser
fix(ui): Correct dropdownControl to style child MenuItem's (#17359)
false
Correct dropdownControl to style child MenuItem's (#17359)
fix
diff --git a/src/sentry/static/sentry/app/components/dropdownControl.jsx b/src/sentry/static/sentry/app/components/dropdownControl.jsx index ae26d3332629d2..16187d52b714c5 100644 --- a/src/sentry/static/sentry/app/components/dropdownControl.jsx +++ b/src/sentry/static/sentry/app/components/dropdownControl.jsx @@ -67,24 +67,22 @@ class DropdownControl extends React.Component { return ( <Container> <DropdownMenu alwaysRenderMenu={alwaysRenderMenu}> - {({isOpen, getMenuProps, getActorProps}) => { - return ( - <React.Fragment> - {this.renderButton(isOpen, getActorProps)} - <MenuContainer - {...getMenuProps()} - alignMenu={alignRight ? 'right' : 'left'} - width={menuWidth} - menuOffset={menuOffset} - isOpen={isOpen} - blendCorner - blendWithActor={blendWithActor} - > - {children} - </MenuContainer> - </React.Fragment> - ); - }} + {({isOpen, getMenuProps, getActorProps}) => ( + <React.Fragment> + {this.renderButton(isOpen, getActorProps)} + <MenuContainer + {...getMenuProps()} + alignMenu={alignRight ? 'right' : 'left'} + width={menuWidth} + menuOffset={menuOffset} + isOpen={isOpen} + blendCorner + blendWithActor={blendWithActor} + > + {children} + </MenuContainer> + </React.Fragment> + )} </DropdownMenu> </Container> ); @@ -112,19 +110,25 @@ const DropdownItem = styled(MenuItem)` font-size: ${p => p.theme.fontSizeMedium}; color: ${p => p.theme.gray2}; - & a { + & a, + & .menu-target { color: ${p => p.theme.foreground}; display: block; padding: ${space(0.5)} ${space(2)}; } - & a:hover { + & a:hover, + & .menu-target:hover { background: ${p => p.theme.offWhite}; } - & a:focus { + & a:focus, + & .menu-target:focus { outline: none; } + &.active a, - &.active a:hover { + &.active a:hover, + &.active .menu-target, + &.active .menu-target:hover { color: ${p => p.theme.white}; background: ${p => p.theme.purple}; }
4ec5694c24cc3b85f66f84980ce8843ff3d7e246
2024-08-16 04:03:15
colin-sentry
feat(eap): Add a method that powers metrics explorer with EAP (#76208)
false
Add a method that powers metrics explorer with EAP (#76208)
feat
diff --git a/requirements-base.txt b/requirements-base.txt index 5fb30f943ac34b..0f41f4b3b3071d 100644 --- a/requirements-base.txt +++ b/requirements-base.txt @@ -67,6 +67,7 @@ rfc3986-validator>=0.1.1 sentry-arroyo>=2.16.5 sentry-kafka-schemas>=0.1.106 sentry-ophio==0.2.7 +sentry-protos>=0.1.3 sentry-redis-tools>=0.1.7 sentry-relay>=0.9.1 sentry-sdk>=2.12.0 diff --git a/requirements-dev-frozen.txt b/requirements-dev-frozen.txt index 48fa12f80ecb71..e38ed492f24d1d 100644 --- a/requirements-dev-frozen.txt +++ b/requirements-dev-frozen.txt @@ -69,6 +69,7 @@ google-resumable-media==2.7.0 googleapis-common-protos==1.63.2 grpc-google-iam-v1==0.13.1 grpc-interceptor==0.15.4 +grpc-stubs==1.53.0.5 grpcio==1.60.1 grpcio-status==1.60.1 h11==0.13.0 @@ -184,6 +185,7 @@ sentry-forked-django-stubs==5.0.4.post1 sentry-forked-djangorestframework-stubs==3.15.0.post1 sentry-kafka-schemas==0.1.106 sentry-ophio==0.2.7 +sentry-protos==0.1.3 sentry-redis-tools==0.1.7 sentry-relay==0.9.1 sentry-sdk==2.12.0 diff --git a/requirements-frozen.txt b/requirements-frozen.txt index 88f4550e997835..4532dc134bed7b 100644 --- a/requirements-frozen.txt +++ b/requirements-frozen.txt @@ -57,6 +57,7 @@ google-resumable-media==2.7.0 googleapis-common-protos==1.63.2 grpc-google-iam-v1==0.13.1 grpc-interceptor==0.15.4 +grpc-stubs==1.53.0.5 grpcio==1.60.1 grpcio-status==1.60.1 h11==0.14.0 @@ -125,6 +126,7 @@ s3transfer==0.10.0 sentry-arroyo==2.16.5 sentry-kafka-schemas==0.1.106 sentry-ophio==0.2.7 +sentry-protos==0.1.3 sentry-redis-tools==0.1.7 sentry-relay==0.9.1 sentry-sdk==2.12.0 diff --git a/src/sentry/api/endpoints/organization_metrics_details.py b/src/sentry/api/endpoints/organization_metrics_details.py index 6293ba0d272136..c1ec211fe1633f 100644 --- a/src/sentry/api/endpoints/organization_metrics_details.py +++ b/src/sentry/api/endpoints/organization_metrics_details.py @@ -1,6 +1,7 @@ from rest_framework.request import Request from rest_framework.response import Response +from sentry import features from sentry.api.api_owners import ApiOwner from sentry.api.api_publish_status import ApiPublishStatus from sentry.api.base import region_silo_endpoint @@ -27,6 +28,24 @@ def get(self, request: Request, organization: Organization) -> Response: {"detail": "You must supply at least one project to see its metrics"}, status=404 ) + if all( + features.has("projects:use-eap-spans-for-metrics-explorer", project) + for project in projects + ): + return Response( + [ + { + "type": "d", + "name": "measurement", + "unit": "none", + "mri": "d:eap/measurement@none", + "operations": ["sum", "avg", "p50", "p95", "p99", "count"], + "projectIds": [project.id for project in projects], + "blockingStatus": [], + } + ] + ) + metrics = get_metrics_meta( organization=organization, projects=projects, use_case_ids=get_use_case_ids(request) ) diff --git a/src/sentry/api/endpoints/organization_metrics_query.py b/src/sentry/api/endpoints/organization_metrics_query.py index ddc84b46f51df3..6c37ffeab4485e 100644 --- a/src/sentry/api/endpoints/organization_metrics_query.py +++ b/src/sentry/api/endpoints/organization_metrics_query.py @@ -4,7 +4,7 @@ from rest_framework.request import Request from rest_framework.response import Response -from sentry import options +from sentry import features, options from sentry.api.api_owners import ApiOwner from sentry.api.api_publish_status import ApiPublishStatus from sentry.api.base import region_silo_endpoint @@ -17,6 +17,7 @@ MQLQuery, run_queries, ) +from sentry.sentry_metrics.querying.eap import mql_eap_bridge from sentry.sentry_metrics.querying.errors import ( InvalidMetricsQueryError, LatestReleaseNotFoundError, @@ -156,6 +157,7 @@ def post(self, request: Request, organization: Organization) -> Response: start, end = get_date_range_from_params(request.GET) interval = self._interval_from_request(request) mql_queries = self._mql_queries_from_request(request) + projects = self.get_projects(request, organization) metrics.incr( key="ddm.metrics_api.query", @@ -166,13 +168,31 @@ def post(self, request: Request, organization: Organization) -> Response: }, ) + if all( + features.has("projects:use-eap-spans-for-metrics-explorer", project) + for project in projects + ): + if len(mql_queries) == 1 and "a" in mql_queries[0].sub_queries: + subquery = mql_queries[0].sub_queries["a"] + if "d:eap/" in subquery.mql: + res_data = mql_eap_bridge.make_eap_request( + subquery.mql, + start, + end, + interval, + organization, + projects, + Referrer.API_ORGANIZATION_METRICS_EAP_QUERY.value, + ) + return Response(status=200, data=res_data) + results = run_queries( mql_queries=mql_queries, start=start, end=end, interval=interval, organization=organization, - projects=self.get_projects(request, organization), + projects=projects, environments=self.get_environments(request, organization), referrer=Referrer.API_ORGANIZATION_METRICS_QUERY.value, query_type=self._get_query_type_from_request(request), diff --git a/src/sentry/api/endpoints/organization_metrics_tag_details.py b/src/sentry/api/endpoints/organization_metrics_tag_details.py index 263e0d19a78182..58ccb308646115 100644 --- a/src/sentry/api/endpoints/organization_metrics_tag_details.py +++ b/src/sentry/api/endpoints/organization_metrics_tag_details.py @@ -2,6 +2,7 @@ from rest_framework.request import Request from rest_framework.response import Response +from sentry import features from sentry.api.api_owners import ApiOwner from sentry.api.api_publish_status import ApiPublishStatus from sentry.api.base import region_silo_endpoint @@ -35,6 +36,28 @@ def get(self, request: Request, organization: Organization, tag_name: str) -> Re {"detail": "You must supply at least one project to see its metrics"}, status=404 ) + if all( + features.has("projects:use-eap-spans-for-metrics-explorer", project) + for project in projects + ): + if len(metric_names) == 1 and metric_names[0].startswith("d:eap"): + # TODO hack for EAP, hardcode some metric names + if tag_name == "color": + return Response( + [ + {"key": tag_name, "value": "red"}, + {"key": tag_name, "value": "blue"}, + {"key": tag_name, "value": "green"}, + ] + ) + if tag_name == "location": + return Response( + [ + {"key": tag_name, "value": "mobile"}, + {"key": tag_name, "value": "frontend"}, + {"key": tag_name, "value": "backend"}, + ] + ) try: mris = convert_metric_names_to_mris(metric_names) tag_values: set[str] = set() diff --git a/src/sentry/api/endpoints/organization_metrics_tags.py b/src/sentry/api/endpoints/organization_metrics_tags.py index 49d156cb0be760..041eb403727ac8 100644 --- a/src/sentry/api/endpoints/organization_metrics_tags.py +++ b/src/sentry/api/endpoints/organization_metrics_tags.py @@ -5,6 +5,7 @@ from rest_framework.request import Request from rest_framework.response import Response +from sentry import features from sentry.api.api_owners import ApiOwner from sentry.api.api_publish_status import ApiPublishStatus from sentry.api.base import region_silo_endpoint @@ -52,6 +53,14 @@ def get(self, request: Request, organization: Organization) -> Response: if not is_mri(metric_name): raise BadRequest(message="Please provide a valid MRI to query a metric's tags.") + if all( + features.has("projects:use-eap-spans-for-metrics-explorer", project) + for project in projects + ): + if metric_name.startswith("d:eap"): + # TODO hack for EAP, return a fixed list + return Response([Tag(key="color"), Tag(key="location")]) + try: if metric_name.startswith("e:"): # If metric_name starts with "e:", and therefore is a derived metric, use the old get_all_tags functionality diff --git a/src/sentry/features/temporary.py b/src/sentry/features/temporary.py index 183905a1391523..316d9141f7c002 100644 --- a/src/sentry/features/temporary.py +++ b/src/sentry/features/temporary.py @@ -539,6 +539,8 @@ def register_temporary_features(manager: FeatureManager): manager.add("projects:span-metrics-extraction", ProjectFeature, FeatureHandlerStrategy.INTERNAL, api_expose=True) manager.add("projects:span-metrics-extraction-addons", ProjectFeature, FeatureHandlerStrategy.INTERNAL, api_expose=False) manager.add("projects:relay-otel-endpoint", ProjectFeature, FeatureHandlerStrategy.OPTIONS, api_expose=False) + # EAP: extremely experimental flag that makes DDM page use EAP tables + manager.add("projects:use-eap-spans-for-metrics-explorer", ProjectFeature, FeatureHandlerStrategy.FLAGPOLE, api_expose=False) # Project plugin features manager.add("projects:plugins", ProjectPluginFeature, FeatureHandlerStrategy.INTERNAL, default=True, api_expose=True) diff --git a/src/sentry/sentry_metrics/querying/eap/README.md b/src/sentry/sentry_metrics/querying/eap/README.md new file mode 100644 index 00000000000000..e12863f9b96068 --- /dev/null +++ b/src/sentry/sentry_metrics/querying/eap/README.md @@ -0,0 +1,5 @@ +We would like to move metrics querying to a span-based system backed by `eap_spans`, part of the Events Analytics Platform work. + +This module facilitates some hacky initial MQL -> GRPC logic, used as a POC for those efforts. + +You should not consider this to be production-ready yet. diff --git a/src/sentry/sentry_metrics/querying/eap/__init__.py b/src/sentry/sentry_metrics/querying/eap/__init__.py new file mode 100644 index 00000000000000..e69de29bb2d1d6 diff --git a/src/sentry/sentry_metrics/querying/eap/mql_eap_bridge.py b/src/sentry/sentry_metrics/querying/eap/mql_eap_bridge.py new file mode 100644 index 00000000000000..9de12630138266 --- /dev/null +++ b/src/sentry/sentry_metrics/querying/eap/mql_eap_bridge.py @@ -0,0 +1,129 @@ +from collections.abc import Iterable, Sequence +from datetime import datetime, timedelta + +import requests +import snuba_sdk.mql.mql +from django.conf import settings +from google.protobuf.timestamp_pb2 import Timestamp as ProtobufTimestamp +from sentry_protos.snuba.v1alpha.endpoint_aggregate_bucket_pb2 import ( + AggregateBucketRequest, + AggregateBucketResponse, +) +from sentry_protos.snuba.v1alpha.request_common_pb2 import RequestMeta +from sentry_protos.snuba.v1alpha.trace_item_filter_pb2 import ( + AndFilter, + OrFilter, + StringFilter, + TraceItemFilter, +) +from snuba_sdk import Timeseries +from snuba_sdk.conditions import And as MQLAnd +from snuba_sdk.conditions import Condition as MQLCondition +from snuba_sdk.conditions import ConditionGroup +from snuba_sdk.conditions import Op as MQLOp +from snuba_sdk.conditions import Or as MQLOr + +from sentry.models.organization import Organization +from sentry.models.project import Project + + +def parse_mql_filters(group: ConditionGroup) -> Iterable[TraceItemFilter]: + for cond in group: + if isinstance(cond, MQLAnd): + yield TraceItemFilter( + and_filter=AndFilter(filters=list(parse_mql_filters(cond.conditions))) + ) + elif isinstance(cond, MQLOr): + yield TraceItemFilter( + or_filter=OrFilter(filters=list(parse_mql_filters(cond.conditions))) + ) + elif isinstance(cond, MQLCondition): + if cond.op == MQLOp.EQ: + yield TraceItemFilter(string_filter=StringFilter(key=cond.lhs.name, value=cond.rhs)) + # TODO: maybe we want to implement other stuff + + +def make_eap_request( + query_mql: str, + start: datetime, + end: datetime, + interval: int, + organization: Organization, + projects: Sequence[Project], + referrer: str, +) -> dict: + start_time_proto = ProtobufTimestamp() + start_time_proto.FromDatetime(start) + end_time_proto = ProtobufTimestamp() + end_time_proto.FromDatetime(end) + + ts: Timeseries = snuba_sdk.mql.mql.parse_mql(query_mql) + + aggregate_map = { + "sum": AggregateBucketRequest.FUNCTION_SUM, + "avg": AggregateBucketRequest.FUNCTION_AVG, + "p50": AggregateBucketRequest.FUNCTION_P50, + "p95": AggregateBucketRequest.FUNCTION_P95, + "P99": AggregateBucketRequest.FUNCTION_P99, + "count": AggregateBucketRequest.FUNCTION_COUNT, + } + + rpc_filters = None + if ts.filters is not None: + rpc_filters = TraceItemFilter( + and_filter=AndFilter(filters=list(parse_mql_filters(ts.filters))) + ) + req = AggregateBucketRequest( + meta=RequestMeta( + organization_id=organization.id, + cogs_category="eap", + referrer=referrer, + project_ids=[project.id for project in projects], + ), + start_timestamp=start_time_proto, + end_timestamp=end_time_proto, + aggregate=aggregate_map[ts.aggregate], + filter=rpc_filters, + ) + http_resp = requests.post(f"{settings.SENTRY_SNUBA}/timeseries", data=req.SerializeToString()) + http_resp.raise_for_status() + + resp = AggregateBucketResponse() + resp.ParseFromString(http_resp.content) + + series_data = list(resp.result) + duration = end - start + bucket_size_secs = duration.total_seconds() / len(series_data) + intervals = [] + for i in range(len(series_data)): + intervals.append((start + timedelta(seconds=bucket_size_secs * i)).isoformat()) + intervals.append(end.isoformat()) + + return { + "data": [ + [ + { + "by": {}, + "totals": None, + "series": series_data, + } + ] + ], + "meta": [ + [ + {"name": "aggregate_value", "type": "Float64"}, + { + "group_bys": [], + "order": "DESC", + "limit": 770, + "has_more": False, + "unit_family": None, + "unit": "none", + "scaling_factor": 1, + }, + ] + ], + "start": start.isoformat(), + "end": end.isoformat(), + "intervals": intervals, + } diff --git a/src/sentry/snuba/referrer.py b/src/sentry/snuba/referrer.py index 16413d970fbdbf..c5408fe29b7ba3 100644 --- a/src/sentry/snuba/referrer.py +++ b/src/sentry/snuba/referrer.py @@ -164,6 +164,7 @@ class Referrer(Enum): "api.organization.metrics-metadata.fetch-metrics-summaries" ) API_ORGANIZATION_METRICS_QUERY = "api.organization.metrics-query" + API_ORGANIZATION_METRICS_EAP_QUERY = "api.organization.metrics-eap-query" API_ORGANIZATION_METRICS_SAMPLES = "api.organization.metrics-samples" API_ORGANIZATION_ISSUE_REPLAY_COUNT = "api.organization-issue-replay-count" API_ORGANIZATION_SDK_UPDATES = "api.organization-sdk-updates"
e0ec88e73008e4baf23d4c8ed64fd53bf6fac486
2020-03-21 00:18:38
Alberto Leal
fix(perf-view): Display total number of transactions on the landing page (#17813)
false
Display total number of transactions on the landing page (#17813)
fix
diff --git a/src/sentry/static/sentry/app/views/eventsV2/chartFooter.tsx b/src/sentry/static/sentry/app/views/eventsV2/chartFooter.tsx index c0e04b15883411..aaeb7b4014ba8d 100644 --- a/src/sentry/static/sentry/app/views/eventsV2/chartFooter.tsx +++ b/src/sentry/static/sentry/app/views/eventsV2/chartFooter.tsx @@ -16,7 +16,7 @@ type Props = { export default function ChartFooter({total, yAxisValue, yAxisOptions, onChange}: Props) { const elements: React.ReactNode[] = []; - elements.push(<SectionHeading key="total-label">{t('Total')}</SectionHeading>); + elements.push(<SectionHeading key="total-label">{t('Total Events')}</SectionHeading>); elements.push( total === null ? ( <SectionValue data-test-id="loading-placeholder" key="total-value"> diff --git a/src/sentry/static/sentry/app/views/performance/charts/footer.tsx b/src/sentry/static/sentry/app/views/performance/charts/footer.tsx index d952cf53a8f721..f981bfd547c783 100644 --- a/src/sentry/static/sentry/app/views/performance/charts/footer.tsx +++ b/src/sentry/static/sentry/app/views/performance/charts/footer.tsx @@ -1,20 +1,93 @@ import React from 'react'; +import * as Sentry from '@sentry/browser'; +import {Location} from 'history'; import {t} from 'app/locale'; +import {Client} from 'app/api'; +import {fetchTotalCount} from 'app/views/eventsV2/utils'; +import EventView, {isAPIPayloadSimilar} from 'app/views/eventsV2/eventView'; +import {Organization} from 'app/types'; -import {ChartControls, SectionHeading} from './styles'; +import {ChartControls, SectionHeading, SectionValue} from './styles'; type Props = { - totals: number | null; + api: Client; + eventView: EventView; + organization: Organization; + location: Location; }; -export default function ChartFooter({totals}: Props) { - return ( - <ChartControls> - <SectionHeading> - {t('Total Events')} - {totals} - </SectionHeading> - </ChartControls> - ); +type State = { + totalValues: null | number; +}; + +class ChartFooter extends React.Component<Props, State> { + state: State = { + totalValues: null, + }; + + componentDidMount() { + this.mounted = true; + + this.fetchTotalCount(); + } + + componentDidUpdate(prevProps: Props) { + const orgSlugHasChanged = + this.props.organization.slug !== prevProps.organization.slug; + const shouldRefetch = this.shouldRefetchData(prevProps); + + if ((orgSlugHasChanged || shouldRefetch) && this.props.eventView.isValid()) { + this.fetchTotalCount(); + } + } + + componentWillUnmount() { + this.mounted = false; + } + + shouldRefetchData = (prevProps: Props): boolean => { + const thisAPIPayload = this.props.eventView.getEventsAPIPayload(this.props.location); + const otherAPIPayload = prevProps.eventView.getEventsAPIPayload(prevProps.location); + + return !isAPIPayloadSimilar(thisAPIPayload, otherAPIPayload); + }; + + mounted: boolean = false; + + async fetchTotalCount() { + const {api, organization, location, eventView} = this.props; + if (!eventView.isValid() || !this.mounted) { + return; + } + + try { + const totals = await fetchTotalCount( + api, + organization.slug, + eventView.getEventsAPIPayload(location) + ); + + if (this.mounted) { + this.setState({totalValues: totals}); + } + } catch (err) { + Sentry.captureException(err); + } + } + + render() { + const {totalValues} = this.state; + + const value = typeof totalValues === 'number' ? totalValues.toLocaleString() : '-'; + + return ( + <ChartControls> + <SectionHeading>{t('Total Events')}</SectionHeading> + <SectionValue>{value}</SectionValue> + </ChartControls> + ); + } } + +export default ChartFooter; diff --git a/src/sentry/static/sentry/app/views/performance/charts/index.tsx b/src/sentry/static/sentry/app/views/performance/charts/index.tsx index 23d68b347a9f86..ffaf8839a593dc 100644 --- a/src/sentry/static/sentry/app/views/performance/charts/index.tsx +++ b/src/sentry/static/sentry/app/views/performance/charts/index.tsx @@ -1,6 +1,5 @@ import React from 'react'; import styled from '@emotion/styled'; -import * as Sentry from '@sentry/browser'; import {Location} from 'history'; import * as ReactRouter from 'react-router'; @@ -13,7 +12,6 @@ import getDynamicText from 'app/utils/getDynamicText'; import {getParams} from 'app/components/organizations/globalSelectionHeader/getParams'; import {Panel} from 'app/components/panels'; import EventView from 'app/views/eventsV2/eventView'; -import {fetchTotalCount} from 'app/views/eventsV2/utils'; import EventsRequest from 'app/views/events/utils/eventsRequest'; import {getUtcToLocalDateObject} from 'app/utils/dates'; import {IconWarning} from 'app/icons'; @@ -36,49 +34,7 @@ type Props = { router: ReactRouter.InjectedRouter; }; -type State = { - totalValues: null | number; -}; - -class Container extends React.Component<Props, State> { - state: State = { - totalValues: null, - }; - - componentDidMount() { - this.mounted = true; - - // TODO: implement later - // this.fetchTotalCount(); - } - - componentWillUnmount() { - this.mounted = false; - } - - mounted: boolean = false; - - async fetchTotalCount() { - const {api, organization, location, eventView} = this.props; - if (!eventView.isValid() || !this.mounted) { - return; - } - - try { - const totals = await fetchTotalCount( - api, - organization.slug, - eventView.getEventsAPIPayload(location) - ); - - if (this.mounted) { - this.setState({totalValues: totals}); - } - } catch (err) { - Sentry.captureException(err); - } - } - +class Container extends React.Component<Props> { render() { const {api, organization, location, eventView, router} = this.props; @@ -158,7 +114,12 @@ class Container extends React.Component<Props, State> { }} </EventsRequest> </ChartsContainer> - <Footer totals={this.state.totalValues} /> + <Footer + api={api} + organization={organization} + eventView={eventView} + location={location} + /> </Panel> ); } diff --git a/src/sentry/static/sentry/app/views/performance/charts/styles.tsx b/src/sentry/static/sentry/app/views/performance/charts/styles.tsx index 95d7a857d119f3..f79d24da4abfce 100644 --- a/src/sentry/static/sentry/app/views/performance/charts/styles.tsx +++ b/src/sentry/static/sentry/app/views/performance/charts/styles.tsx @@ -15,6 +15,12 @@ export const SectionHeading = styled('h4')` line-height: 1.2; `; +export const SectionValue = styled('span')` + color: ${p => p.theme.gray3}; + font-size: ${p => p.theme.fontSizeMedium}; + margin-right: ${space(1)}; +`; + export const ChartsContainer = styled('div')` padding: ${space(2)} ${space(1.5)}; `; @@ -31,7 +37,7 @@ export const ChartContainer = styled('div')` export const ChartControls = styled('div')` display: flex; - justify-content: space-between; + align-items: center; padding: ${space(1)} ${space(3)}; border-top: 1px solid ${p => p.theme.borderLight}; `;
559022034dfd5ecc2c8510c50a4b18fc5c24cb7a
2023-06-24 04:57:54
Gabe Villalobos
ref(HC): Revokes org and org mapping write permissions, adds outbox handling to org model (#51544)
false
Revokes org and org mapping write permissions, adds outbox handling to org model (#51544)
ref
diff --git a/src/sentry/models/organization.py b/src/sentry/models/organization.py index 0701b02b314e17..8f5ec56c9b70b3 100644 --- a/src/sentry/models/organization.py +++ b/src/sentry/models/organization.py @@ -11,6 +11,7 @@ from django.urls import NoReverseMatch, reverse from django.utils import timezone from django.utils.functional import cached_property +from typing_extensions import override from bitfield import BitField from sentry import features, roles @@ -265,6 +266,16 @@ def save(self, *args, **kwargs): with outbox_context(transaction.atomic()): self.save_with_update_outbox(*args, **kwargs) + # Override for the default update method to ensure that most atomic updates + # generate an outbox alongside any mutations to ensure data is replicated + # properly to the control silo. + @override + def update(self, *args, **kwargs): + with outbox_context(transaction.atomic()): + results = super().update(*args, **kwargs) + Organization.outbox_for_update(self.id).save() + return results + @classmethod def reserve_snowflake_id(cls): return generate_snowflake_id(cls.snowflake_redis_key) diff --git a/src/sentry/services/hybrid_cloud/organization/impl.py b/src/sentry/services/hybrid_cloud/organization/impl.py index 6ceda78f023850..48aafd743fd2f9 100644 --- a/src/sentry/services/hybrid_cloud/organization/impl.py +++ b/src/sentry/services/hybrid_cloud/organization/impl.py @@ -30,6 +30,7 @@ from sentry.services.hybrid_cloud.organization import ( OrganizationService, OrganizationSignalService, + RpcOrganization, RpcOrganizationFlagsUpdate, RpcOrganizationInvite, RpcOrganizationMember, @@ -286,7 +287,9 @@ def update_flags(self, *, organization_id: int, flags: RpcOrganizationFlagsUpdat else: raise TypeError(f"Invalid value received for update_flags: {name}={value!r}") - Organization.objects.filter(id=organization_id).update(flags=updates) + with outbox_context(transaction.atomic()): + Organization.objects.filter(id=organization_id).update(flags=updates) + Organization.outbox_for_update(org_id=organization_id).save() @staticmethod def _deserialize_member_flags(flags: RpcOrganizationMemberFlags) -> int: @@ -387,9 +390,7 @@ def get_top_dog_team_member_ids(self, organization_id: int) -> List[int]: ) ) - def update_default_role( - self, *, organization_id: int, default_role: str - ) -> RpcOrganizationMember: + def update_default_role(self, *, organization_id: int, default_role: str) -> RpcOrganization: org = Organization.objects.get(id=organization_id) org.default_role = default_role org.save() diff --git a/src/sentry/services/hybrid_cloud/organization_actions/impl.py b/src/sentry/services/hybrid_cloud/organization_actions/impl.py index e3abb5b38eb898..6099ea6f759bac 100644 --- a/src/sentry/services/hybrid_cloud/organization_actions/impl.py +++ b/src/sentry/services/hybrid_cloud/organization_actions/impl.py @@ -3,7 +3,7 @@ from django.db import transaction from django.db.models.expressions import CombinedExpression -from sentry.models import Organization, OrganizationStatus +from sentry.models import Organization, OrganizationStatus, outbox_context class OrganizationCreateAndUpdateOptions(TypedDict, total=False): @@ -17,42 +17,59 @@ class OrganizationCreateAndUpdateOptions(TypedDict, total=False): def create_organization_with_outbox_message( *, create_options: OrganizationCreateAndUpdateOptions ) -> Organization: - with transaction.atomic(): - org: Organization = Organization.objects.create(**create_options) - Organization.outbox_for_update(org_id=org.id).save() + org: Organization = Organization.objects.create(**create_options) return org def update_organization_with_outbox_message( *, org_id: int, update_data: OrganizationCreateAndUpdateOptions ) -> Organization: - with transaction.atomic(): + with outbox_context(transaction.atomic()): org: Organization = Organization.objects.get(id=org_id) org.update(**update_data) - Organization.outbox_for_update(org_id=org.id).save() - org.refresh_from_db() - return org + org.refresh_from_db() + return org def upsert_organization_by_org_id_with_outbox_message( *, org_id: int, upsert_data: OrganizationCreateAndUpdateOptions ) -> Organization: - with transaction.atomic(): + with outbox_context(transaction.atomic()): org, created = Organization.objects.update_or_create(id=org_id, defaults=upsert_data) - Organization.outbox_for_update(org_id=org_id).save() return org def mark_organization_as_pending_deletion_with_outbox_message( *, org_id: int ) -> Optional[Organization]: - with transaction.atomic(): - query_result = Organization.objects.filter( + with outbox_context(transaction.atomic()): + update_count = Organization.objects.filter( id=org_id, status=OrganizationStatus.ACTIVE ).update(status=OrganizationStatus.PENDING_DELETION) - if not query_result: + if not update_count: + return None + + Organization.outbox_for_update(org_id=org_id).save() + + org = Organization.objects.get(id=org_id) + return org + + +def unmark_organization_as_pending_deletion_with_outbox_message( + *, org_id: int +) -> Optional[Organization]: + with outbox_context(transaction.atomic()): + update_count = Organization.objects.filter( + id=org_id, + status__in=[ + OrganizationStatus.PENDING_DELETION, + OrganizationStatus.DELETION_IN_PROGRESS, + ], + ).update(status=OrganizationStatus.ACTIVE) + + if not update_count: return None Organization.outbox_for_update(org_id=org_id).save() diff --git a/src/sentry/services/hybrid_cloud/organization_mapping/impl.py b/src/sentry/services/hybrid_cloud/organization_mapping/impl.py index fcb847d241bec4..2a10080e65e172 100644 --- a/src/sentry/services/hybrid_cloud/organization_mapping/impl.py +++ b/src/sentry/services/hybrid_cloud/organization_mapping/impl.py @@ -1,5 +1,6 @@ from typing import List, Optional +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.models.organizationmapping import OrganizationMapping from sentry.services.hybrid_cloud.organization_mapping import ( OrganizationMappingService, @@ -68,11 +69,12 @@ def update(self, organization_id: int, update: RpcOrganizationMappingUpdate) -> def upsert( self, organization_id: int, update: RpcOrganizationMappingUpdate ) -> RpcOrganizationMapping: - org_mapping, _created = OrganizationMapping.objects.update_or_create( - organization_id=organization_id, defaults=update - ) + with in_test_psql_role_override("postgres"): + org_mapping, _created = OrganizationMapping.objects.update_or_create( + organization_id=organization_id, defaults=update + ) - return serialize_organization_mapping(org_mapping) + return serialize_organization_mapping(org_mapping) def verify_mappings(self, organization_id: int, slug: str) -> None: try: diff --git a/src/sentry/testutils/helpers/api_gateway.py b/src/sentry/testutils/helpers/api_gateway.py index 37984960e6a79e..b17b97371323a7 100644 --- a/src/sentry/testutils/helpers/api_gateway.py +++ b/src/sentry/testutils/helpers/api_gateway.py @@ -9,6 +9,7 @@ from sentry.api.base import control_silo_endpoint, region_silo_endpoint from sentry.api.bases.organization import OrganizationEndpoint +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.models.organizationmapping import OrganizationMapping from sentry.testutils import APITestCase from sentry.types.region import Region, RegionCategory, clear_global_regions @@ -133,9 +134,11 @@ def setUp(self): content_type="application/json", adding_headers={"test": "header"}, ) - OrganizationMapping.objects.get(organization_id=self.organization.id).update( - region_name="region1" - ) + + with in_test_psql_role_override("postgres"): + OrganizationMapping.objects.get(organization_id=self.organization.id).update( + region_name="region1" + ) # Echos the request body and header back for verification def return_request_body(request): diff --git a/src/sentry/utils/migrations.py b/src/sentry/utils/migrations.py index 142f999c425bbf..7a50571c859eaa 100644 --- a/src/sentry/utils/migrations.py +++ b/src/sentry/utils/migrations.py @@ -1,5 +1,6 @@ from django.db.models import F +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.utils.query import RangeQuerySetWrapperWithProgressBar @@ -14,4 +15,5 @@ def clear_flag(Model, flag_name, flag_attr_name="flags"): update_kwargs = { flag_attr_name: F(flag_attr_name).bitand(~getattr(Model, flag_attr_name)[flag_name]) } - Model.objects.filter(id=item.id).update(**update_kwargs) + with in_test_psql_role_override("postgres"): + Model.objects.filter(id=item.id).update(**update_kwargs) diff --git a/src/sentry/web/frontend/restore_organization.py b/src/sentry/web/frontend/restore_organization.py index 8deb336c295828..8c54035446c7f8 100644 --- a/src/sentry/web/frontend/restore_organization.py +++ b/src/sentry/web/frontend/restore_organization.py @@ -8,6 +8,9 @@ from sentry.api import client from sentry.models import Organization, OrganizationStatus from sentry.services.hybrid_cloud.organization import organization_service +from sentry.services.hybrid_cloud.organization_actions.impl import ( + unmark_organization_as_pending_deletion_with_outbox_message, +) from sentry.web.frontend.base import OrganizationView from sentry.web.helpers import render_to_response @@ -65,9 +68,10 @@ def post(self, request: Request, organization) -> Response: messages.add_message(request, messages.ERROR, ERR_MESSAGES[organization.status]) return self.redirect(reverse("sentry")) - updated = Organization.objects.filter( - id=organization.id, status__in=deletion_statuses - ).update(status=OrganizationStatus.ACTIVE) + updated = unmark_organization_as_pending_deletion_with_outbox_message( + org_id=organization.id + ) + if updated: client.put( f"/organizations/{organization.slug}/", diff --git a/tests/conftest.py b/tests/conftest.py index 16b87de74b5770..6378a9bd94f0fa 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -172,7 +172,12 @@ def protect_hybrid_cloud_writes_and_deletes(request): create Outbox objects in the same transaction that matches what you delete. """ from sentry.db.models.fields.hybrid_cloud_foreign_key import HybridCloudForeignKey - from sentry.models import OrganizationMember, OrganizationMemberMapping + from sentry.models import ( + Organization, + OrganizationMapping, + OrganizationMember, + OrganizationMemberMapping, + ) from sentry.testutils.silo import iter_models, reset_test_role, restrict_role try: @@ -205,6 +210,10 @@ def protect_hybrid_cloud_writes_and_deletes(request): # outboxes in a transaction, and cover that transaction with `in_test_psql_role_override` restrict_role(role="postgres_unprivileged", model=OrganizationMember, revocation_type="INSERT") restrict_role(role="postgres_unprivileged", model=OrganizationMember, revocation_type="UPDATE") + restrict_role(role="postgres_unprivileged", model=Organization, revocation_type="INSERT") + restrict_role(role="postgres_unprivileged", model=Organization, revocation_type="UPDATE") + restrict_role(role="postgres_unprivileged", model=OrganizationMapping, revocation_type="INSERT") + restrict_role(role="postgres_unprivileged", model=OrganizationMapping, revocation_type="UPDATE") # OrganizationMember objects need to cascade, but they can't use the standard hybrid cloud foreign key because the # identifiers are not snowflake ids. restrict_role(role="postgres_unprivileged", model=OrganizationMember, revocation_type="DELETE") diff --git a/tests/sentry/api/endpoints/test_accept_organization_invite.py b/tests/sentry/api/endpoints/test_accept_organization_invite.py index 221e89a9775c52..928ffc5161249f 100644 --- a/tests/sentry/api/endpoints/test_accept_organization_invite.py +++ b/tests/sentry/api/endpoints/test_accept_organization_invite.py @@ -156,13 +156,14 @@ def test_multi_region_organizationmember_id(self): ), ] ): - self.create_organization_mapping( - organization_id=101010, - slug="abcslug", - name="The Thing", - idempotency_key="", - region_name="some-region", - ) + with in_test_psql_role_override("postgres"): + self.create_organization_mapping( + organization_id=101010, + slug="abcslug", + name="The Thing", + idempotency_key="", + region_name="some-region", + ) self._require_2fa_for_organization() assert not self.user.has_2fa() diff --git a/tests/sentry/hybrid_cloud/test_organizationmapping.py b/tests/sentry/hybrid_cloud/test_organizationmapping.py index b9d1edefeafbdc..00f014f8a7f047 100644 --- a/tests/sentry/hybrid_cloud/test_organizationmapping.py +++ b/tests/sentry/hybrid_cloud/test_organizationmapping.py @@ -1,7 +1,7 @@ import pytest from django.db import IntegrityError -from sentry.models import Organization, outbox_context +from sentry.models import outbox_context from sentry.models.organization import OrganizationStatus from sentry.models.organizationmapping import OrganizationMapping from sentry.services.hybrid_cloud.organization_mapping import ( @@ -9,7 +9,6 @@ organization_mapping_service, ) from sentry.testutils import TransactionTestCase -from sentry.testutils.factories import Factories from sentry.testutils.outbox import outbox_runner from sentry.testutils.silo import control_silo_test, exempt_from_silo_limits @@ -18,10 +17,9 @@ class OrganizationMappingTest(TransactionTestCase): def test_create_on_organization_save(self): with outbox_context(flush=False), exempt_from_silo_limits(): - organization = Organization( + organization = self.create_organization( name="test name", ) - organization.save() # Validate that organization mapping has not been created with pytest.raises(OrganizationMapping.DoesNotExist): @@ -64,16 +62,11 @@ def test_upsert__create_if_not_found(self): def test_upsert__update_if_found(self): with exempt_from_silo_limits(): - self.organization = Organization( + self.organization = self.create_organization( name="test name", slug="foobar", ) - self.organization.save() - - with outbox_runner(): - pass - fixture_org_mapping = OrganizationMapping.objects.get(organization_id=self.organization.id) organization_mapping_service.upsert( @@ -89,7 +82,7 @@ def test_upsert__update_if_found(self): assert fixture_org_mapping.status == OrganizationStatus.PENDING_DELETION def test_upsert__duplicate_slug(self): - self.organization = Factories.create_organization(slug="alreadytaken") + self.organization = self.create_organization(slug="alreadytaken") with pytest.raises(IntegrityError): organization_mapping_service.upsert( diff --git a/tests/sentry/hybrid_cloud/test_region.py b/tests/sentry/hybrid_cloud/test_region.py index ce246db091a897..f70300b04fe748 100644 --- a/tests/sentry/hybrid_cloud/test_region.py +++ b/tests/sentry/hybrid_cloud/test_region.py @@ -1,5 +1,6 @@ import pytest +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.models.organizationmapping import OrganizationMapping from sentry.models.organizationmember import OrganizationMember from sentry.services.hybrid_cloud.region import ( @@ -26,8 +27,9 @@ def setUp(self): self.target_region = self.regions[0] self.organization = self.create_organization() org_mapping = OrganizationMapping.objects.get(organization_id=self.organization.id) - org_mapping.region_name = self.target_region.name - org_mapping.save() + with in_test_psql_role_override("postgres"): + org_mapping.region_name = self.target_region.name + org_mapping.save() def test_by_organization_object(self): with override_regions(self.regions): diff --git a/tests/sentry/hybrid_cloud/test_rpc.py b/tests/sentry/hybrid_cloud/test_rpc.py index d3e7e67d773ba8..e66245b27984e1 100644 --- a/tests/sentry/hybrid_cloud/test_rpc.py +++ b/tests/sentry/hybrid_cloud/test_rpc.py @@ -4,6 +4,7 @@ import pytest from django.test import override_settings +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.models import OrganizationMapping from sentry.services.hybrid_cloud.actor import RpcActor from sentry.services.hybrid_cloud.auth import AuthService @@ -45,14 +46,15 @@ def test_remote_service(self, mock_dispatch_remote_call): user = self.create_user() organization = self.create_organization() - OrganizationMapping.objects.update_or_create( - organization_id=organization.id, - defaults={ - "slug": organization.slug, - "name": organization.name, - "region_name": target_region.name, - }, - ) + with in_test_psql_role_override("postgres"): + OrganizationMapping.objects.update_or_create( + organization_id=organization.id, + defaults={ + "slug": organization.slug, + "name": organization.name, + "region_name": target_region.name, + }, + ) serial_user = RpcUser(id=user.id) serial_org = serialize_rpc_organization(organization) diff --git a/tests/sentry/integrations/test_pipeline.py b/tests/sentry/integrations/test_pipeline.py index 45bffb59e75fb6..e5b4135f96a396 100644 --- a/tests/sentry/integrations/test_pipeline.py +++ b/tests/sentry/integrations/test_pipeline.py @@ -1,6 +1,7 @@ from unittest.mock import patch from sentry.api.utils import generate_organization_url +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.integrations.example import AliasedIntegrationProvider, ExampleIntegrationProvider from sentry.integrations.gitlab.integration import GitlabIntegrationProvider from sentry.models import ( @@ -57,7 +58,7 @@ def _setup_region_restriction(self): integration = Integration.objects.create( name="test", external_id=self.external_id, provider=self.provider.key ) - with receivers_raise_on_send(), outbox_runner(): + with receivers_raise_on_send(), outbox_runner(), in_test_psql_role_override("postgres"): for org in na_orgs: integration.add_organization(org) mapping = OrganizationMapping.objects.get(organization_id=org.id) @@ -137,7 +138,9 @@ def test_is_violating_region_restriction_success(self, *args): # Installing organization is from the same region mapping = OrganizationMapping.objects.get(organization_id=self.organization.id) - mapping.update(region_name="na") + + with in_test_psql_role_override("postgres"): + mapping.update(region_name="na") self.pipeline.state.data = {"external_id": self.external_id} with patch("sentry.integrations.pipeline.IntegrationPipeline._dialog_response") as resp: @@ -152,7 +155,9 @@ def test_is_violating_region_restriction_failure(self, *args): # Installing organization is from a different region mapping = OrganizationMapping.objects.get(organization_id=self.organization.id) - mapping.update(region_name="eu") + + with in_test_psql_role_override("postgres"): + mapping.update(region_name="eu") self.pipeline.state.data = {"external_id": self.external_id} with patch("sentry.integrations.pipeline.IntegrationPipeline._dialog_response") as resp: diff --git a/tests/sentry/runner/commands/test_backup.py b/tests/sentry/runner/commands/test_backup.py index 151b1d17d60f89..95dd57140dde79 100644 --- a/tests/sentry/runner/commands/test_backup.py +++ b/tests/sentry/runner/commands/test_backup.py @@ -2,6 +2,7 @@ from click.testing import CliRunner from django.db import IntegrityError +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.runner.commands.backup import export, import_ from sentry.utils import json @@ -16,7 +17,8 @@ def backup_json_filename(tmp_path): @pytest.mark.django_db def test_import(backup_json_filename): - rv = CliRunner().invoke(import_, backup_json_filename) + with in_test_psql_role_override("postgres"): + rv = CliRunner().invoke(import_, backup_json_filename) assert rv.exit_code == 0, rv.output @@ -31,7 +33,9 @@ def test_import_duplicate_key(backup_json_filename): contents.append(duplicate_key_item) with open(backup_json_filename, "w") as backup_file: backup_file.write(json.dumps(contents)) - rv = CliRunner().invoke(import_, backup_json_filename) + + with in_test_psql_role_override("postgres"): + rv = CliRunner().invoke(import_, backup_json_filename) assert ( rv.output == ">> Are you restoring from a backup of the same version of Sentry?\n>> Are you restoring onto a clean database?\n>> If so then this IntegrityError might be our fault, you can open an issue here:\n>> https://github.com/getsentry/sentry/issues/new/choose\n" diff --git a/tests/sentry/services/test_organization_actions.py b/tests/sentry/services/test_organization_actions.py index 2f90af20edc043..399e5723fe3a4a 100644 --- a/tests/sentry/services/test_organization_actions.py +++ b/tests/sentry/services/test_organization_actions.py @@ -11,11 +11,11 @@ from sentry.services.hybrid_cloud.organization_actions.impl import ( create_organization_with_outbox_message, mark_organization_as_pending_deletion_with_outbox_message, + unmark_organization_as_pending_deletion_with_outbox_message, update_organization_with_outbox_message, upsert_organization_by_org_id_with_outbox_message, ) from sentry.testutils import TestCase -from sentry.testutils.outbox import outbox_runner from sentry.testutils.silo import region_silo_test @@ -37,13 +37,7 @@ class OrganizationUpdateTest(TestCase): def setUp(self): self.org: Organization = self.create_organization(slug="sluggy", name="barfoo") - with outbox_runner(): - pass - def test_create_organization_with_outbox_message(self): - with outbox_runner(): - pass - with outbox_context(flush=False): org: Organization = create_organization_with_outbox_message( create_options={ @@ -56,7 +50,7 @@ def test_create_organization_with_outbox_message(self): assert org.id assert org.slug == "santry" assert org.name == "santry" - assert_outbox_update_message_exists(org=org, expected_count=2) + assert_outbox_update_message_exists(org=org, expected_count=1) @region_silo_test(stable=True) @@ -64,11 +58,11 @@ class OrganizationUpdateWithOutboxTest(TestCase): def setUp(self): self.org: Organization = self.create_organization(slug="sluggy", name="barfoo") - with outbox_runner(): - pass - def test_update_organization_with_outbox_message(self): - update_organization_with_outbox_message(org_id=self.org.id, update_data={"name": "foobar"}) + with outbox_context(flush=False): + update_organization_with_outbox_message( + org_id=self.org.id, update_data={"name": "foobar"} + ) self.org.refresh_from_db() assert self.org.name == "foobar" @@ -85,9 +79,6 @@ class OrganizationUpsertWithOutboxTest(TestCase): def setUp(self): self.org: Organization = self.create_organization(slug="sluggy", name="barfoo") - with outbox_runner(): - pass - def test_upsert_queues_outbox_message_and_updates_org(self): # The test fixture creates at least 1 org so comparing count before # and after the upsert is the safest way to assert we haven't created @@ -116,7 +107,7 @@ def test_upsert_queues_outbox_message_and_updates_org(self): == org_before_modification.default_role ) - assert_outbox_update_message_exists(org=self.org, expected_count=2) + assert_outbox_update_message_exists(org=self.org, expected_count=1) def test_upsert_creates_organization_with_desired_id(self): previous_org_count = Organization.objects.count() @@ -143,7 +134,7 @@ def test_upsert_creates_organization_with_desired_id(self): assert org_before_modification.slug == self.org.slug assert org_before_modification.name == self.org.name assert org_before_modification.status == self.org.status - assert_outbox_update_message_exists(org=db_created_org, expected_count=2) + assert_outbox_update_message_exists(org=db_created_org, expected_count=1) @region_silo_test(stable=True) @@ -155,7 +146,11 @@ def setUp(self): def test_mark_for_deletion_and_outbox_generation(self): org_before_update = Organization.objects.get(id=self.org.id) - updated_org = mark_organization_as_pending_deletion_with_outbox_message(org_id=self.org.id) + + with outbox_context(flush=False): + updated_org = mark_organization_as_pending_deletion_with_outbox_message( + org_id=self.org.id + ) assert updated_org self.org.refresh_from_db() @@ -167,12 +162,14 @@ def test_mark_for_deletion_and_outbox_generation(self): def test_mark_for_deletion_on_already_deleted_org(self): self.org.status = OrganizationStatus.PENDING_DELETION - with outbox_runner(): - self.org.save() + self.org.save() org_before_update = Organization.objects.get(id=self.org.id) - updated_org = mark_organization_as_pending_deletion_with_outbox_message(org_id=self.org.id) + with outbox_context(flush=False): + updated_org = mark_organization_as_pending_deletion_with_outbox_message( + org_id=self.org.id + ) assert updated_org is None @@ -182,3 +179,65 @@ def test_mark_for_deletion_on_already_deleted_org(self): assert self.org.slug == org_before_update.slug assert_outbox_update_message_exists(self.org, 0) + + +@region_silo_test(stable=True) +class UnmarkOrganizationForDeletionWithOutboxMessageTest(TestCase): + def setUp(self): + self.org: Organization = self.create_organization( + slug="sluggy", name="barfoo", status=OrganizationStatus.PENDING_DELETION + ) + + def test_unmark_for_pending_deletion_and_outbox_generation(self): + with outbox_context(flush=False): + updated_org = unmark_organization_as_pending_deletion_with_outbox_message( + org_id=self.org.id + ) + + assert updated_org + self.org.refresh_from_db() + + assert updated_org.status == self.org.status == OrganizationStatus.ACTIVE + assert updated_org.name == self.org.name + assert updated_org.slug == self.org.slug + + assert_outbox_update_message_exists(self.org, 1) + + def test_unmark_for_deletion_in_progress_and_outbox_generation(self): + update_organization_with_outbox_message( + org_id=self.org.id, update_data={"status": OrganizationStatus.DELETION_IN_PROGRESS} + ) + + with outbox_context(flush=False): + updated_org = unmark_organization_as_pending_deletion_with_outbox_message( + org_id=self.org.id + ) + + assert updated_org + self.org.refresh_from_db() + + assert updated_org.status == self.org.status == OrganizationStatus.ACTIVE + assert updated_org.name == self.org.name + assert updated_org.slug == self.org.slug + + assert_outbox_update_message_exists(self.org, 1) + + def test_unmark_org_when_already_active(self): + update_organization_with_outbox_message( + org_id=self.org.id, update_data={"status": OrganizationStatus.ACTIVE} + ) + + org_before_update = Organization.objects.get(id=self.org.id) + + with outbox_context(flush=False): + updated_org = unmark_organization_as_pending_deletion_with_outbox_message( + org_id=self.org.id + ) + + assert not updated_org + + self.org.refresh_from_db() + assert self.org.status == org_before_update.status + assert self.org.name == org_before_update.name + assert self.org.slug == org_before_update.slug + assert_outbox_update_message_exists(self.org, 0) diff --git a/tests/sentry/tasks/test_organization_mapping.py b/tests/sentry/tasks/test_organization_mapping.py index d1ff2e3745be82..c8a4c7e7878363 100644 --- a/tests/sentry/tasks/test_organization_mapping.py +++ b/tests/sentry/tasks/test_organization_mapping.py @@ -2,6 +2,7 @@ import pytest +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.models.organization import Organization from sentry.models.organizationmapping import OrganizationMapping from sentry.tasks.organization_mapping import ORGANIZATION_MAPPING_EXPIRY, repair_mappings @@ -9,6 +10,12 @@ from sentry.testutils.factories import Factories [email protected](autouse=True) +def role_override(): + with in_test_psql_role_override("postgres"): + yield + + class OrganizationMappingRepairTest(TestCase): def test_removes_expired_unverified(self): self.organization = Factories.create_organization() @@ -17,6 +24,7 @@ def test_removes_expired_unverified(self): mapping.verified = False mapping.date_created = expired_time mapping.save() + phantom_mapping = self.create_organization_mapping( Organization(id=123, slug="fake-slug"), date_created=expired_time, verified=False ) diff --git a/tests/sentry/types/test_region.py b/tests/sentry/types/test_region.py index 52b9b47a7fd01d..0cf19ca9863625 100644 --- a/tests/sentry/types/test_region.py +++ b/tests/sentry/types/test_region.py @@ -2,6 +2,7 @@ from django.conf import settings from django.test import override_settings +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.models import OrganizationMapping from sentry.services.hybrid_cloud.organization import organization_service from sentry.silo import SiloMode @@ -56,7 +57,7 @@ def test_get_region_for_organization(self): Region("eu", 2, "http://eu.testserver", RegionCategory.MULTI_TENANT), ] mapping = OrganizationMapping.objects.get(slug=self.organization.slug) - with override_regions(regions): + with override_regions(regions), in_test_psql_role_override("postgres"): mapping.update(region_name="az") with pytest.raises(RegionResolutionError): # Region does not exist @@ -104,7 +105,9 @@ def test_find_regions_for_user(self): organization_mapping.name = "test name" organization_mapping.region_name = "na" organization_mapping.idempotency_key = "test" - organization_mapping.save() + + with in_test_psql_role_override("postgres"): + organization_mapping.save() region_config = [ { diff --git a/tests/sentry/utils/test_audit.py b/tests/sentry/utils/test_audit.py index 7809ed60ac0ad1..377d5dcd46d174 100644 --- a/tests/sentry/utils/test_audit.py +++ b/tests/sentry/utils/test_audit.py @@ -1,6 +1,7 @@ from django.contrib.auth.models import AnonymousUser from sentry import audit_log +from sentry.db.postgres.roles import in_test_psql_role_override from sentry.models import ( AuditLogEntry, DeletedOrganization, @@ -90,7 +91,7 @@ def test_audit_entry_org_delete_log(self): self.assert_valid_deleted_log(deleted_org, self.org) def test_audit_entry_org_restore_log(self): - with exempt_from_silo_limits(): + with exempt_from_silo_limits(), in_test_psql_role_override("postgres"): Organization.objects.filter(id=self.organization.id).update( status=OrganizationStatus.PENDING_DELETION ) diff --git a/tests/sentry/web/frontend/test_auth_organization_login.py b/tests/sentry/web/frontend/test_auth_organization_login.py index 929d00b220823d..0d2c03d7e592f2 100644 --- a/tests/sentry/web/frontend/test_auth_organization_login.py +++ b/tests/sentry/web/frontend/test_auth_organization_login.py @@ -11,7 +11,6 @@ from sentry.models import ( AuthIdentity, AuthProvider, - Organization, OrganizationMember, OrganizationOption, OrganizationStatus, @@ -1055,9 +1054,7 @@ def test_anonymous_user_with_automatic_migration(self): assert resp.status_code == 200 def test_org_not_visible(self): - Organization.objects.filter(id=self.organization.id).update( - status=OrganizationStatus.DELETION_IN_PROGRESS - ) + self.organization.update(status=OrganizationStatus.DELETION_IN_PROGRESS) resp = self.client.get(self.path, follow=True) assert resp.status_code == 200 diff --git a/tests/sentry/web/frontend/test_restore_organization.py b/tests/sentry/web/frontend/test_restore_organization.py index b3204cb5efbb5e..b2b7021b2f639d 100644 --- a/tests/sentry/web/frontend/test_restore_organization.py +++ b/tests/sentry/web/frontend/test_restore_organization.py @@ -48,9 +48,7 @@ def test_renders_with_context(self): assert resp.context["deleting_organization"] == self.organization assert resp.context["pending_deletion"] is True - Organization.objects.filter(id=self.organization.id).update( - status=OrganizationStatus.DELETION_IN_PROGRESS - ) + self.organization.update(status=OrganizationStatus.DELETION_IN_PROGRESS) resp = self.client.get(self.path) @@ -73,9 +71,7 @@ def test_renders_with_context_customer_domain(self): assert resp.context["deleting_organization"] == self.organization assert resp.context["pending_deletion"] is True - Organization.objects.filter(id=self.organization.id).update( - status=OrganizationStatus.DELETION_IN_PROGRESS - ) + self.organization.update(status=OrganizationStatus.DELETION_IN_PROGRESS) resp = self.client.get(path, SERVER_NAME=f"{self.organization.slug}.testserver") @@ -96,9 +92,7 @@ def test_success(self): assert org.status == OrganizationStatus.ACTIVE def test_too_late_still_restores(self): - Organization.objects.filter(id=self.organization.id).update( - status=OrganizationStatus.DELETION_IN_PROGRESS - ) + self.organization.update(status=OrganizationStatus.DELETION_IN_PROGRESS) resp = self.client.post(self.path) @@ -112,7 +106,7 @@ def test_org_already_deleted(self): assert ScheduledDeletion.objects.count() == 0 org_id = self.organization.id - Organization.objects.filter(id=org_id).update(status=OrganizationStatus.PENDING_DELETION) + self.organization.update(status=OrganizationStatus.PENDING_DELETION) deletion = ScheduledDeletion.schedule(self.organization, days=0) deletion.update(in_progress=True)
0587f8718fca61d1cfb024f625d3d2d823fe5366
2024-06-24 16:01:54
ArthurKnaus
feat(metrics): Remove cardinality limit setting (#73169)
false
Remove cardinality limit setting (#73169)
feat
diff --git a/static/app/views/settings/projectMetrics/cardinalityLimit.tsx b/static/app/views/settings/projectMetrics/cardinalityLimit.tsx deleted file mode 100644 index 9c561043e054e3..00000000000000 --- a/static/app/views/settings/projectMetrics/cardinalityLimit.tsx +++ /dev/null @@ -1,77 +0,0 @@ -import styled from '@emotion/styled'; - -import Access from 'sentry/components/acl/access'; -import NumberField from 'sentry/components/forms/fields/numberField'; -import Form from 'sentry/components/forms/form'; -import Panel from 'sentry/components/panels/panel'; -import PanelBody from 'sentry/components/panels/panelBody'; -import PanelHeader from 'sentry/components/panels/panelHeader'; -import {t} from 'sentry/locale'; -import type {Project} from 'sentry/types/project'; -import useOrganization from 'sentry/utils/useOrganization'; - -import {DEFAULT_METRICS_CARDINALITY_LIMIT} from '../../../utils/metrics/constants'; - -type Props = { - project: Project; -}; - -function transformData(data) { - const limit = data.relayCustomMetricCardinalityLimit; - return { - relayCustomMetricCardinalityLimit: - limit === '' || limit === DEFAULT_METRICS_CARDINALITY_LIMIT.toString() - ? null - : limit, - }; -} - -export function CardinalityLimit({project}: Props) { - const organization = useOrganization(); - const endpoint = `/projects/${organization.slug}/${project.slug}/`; - - return ( - <Form - apiEndpoint={endpoint} - apiMethod="PUT" - saveOnBlur - initialData={{ - relayCustomMetricCardinalityLimit: - project.relayCustomMetricCardinalityLimit ?? - DEFAULT_METRICS_CARDINALITY_LIMIT.toString(), - }} - > - <Panel> - <PanelHeader>{t('Limits')}</PanelHeader> - <PanelBody> - <Access access={['project:write']} project={project}> - {({hasAccess}) => ( - <StyledNumberField - disabledReason={ - !hasAccess - ? t('You do not have permission to edit the cardinality limit') - : undefined - } - disabled={!hasAccess} - name="relayCustomMetricCardinalityLimit" - label={t('Cardinality Limit')} - help={t( - 'The cardinality limit defines the maximum number of unique tag combinations allowed for a metric (measured per hour). If the cardinality limit is exceeded, the metric will be blocked.' - )} - saveOnBlur - placeholder={t('Enter a value')} - flexibleControlStateSize - multiple - getData={transformData} - /> - )} - </Access> - </PanelBody> - </Panel> - </Form> - ); -} - -const StyledNumberField = styled(NumberField)` - ${p => p.disabled && `cursor: not-allowed`} -`; diff --git a/static/app/views/settings/projectMetrics/projectMetrics.tsx b/static/app/views/settings/projectMetrics/projectMetrics.tsx index 25ca29e034b9bc..d4f76679765a37 100644 --- a/static/app/views/settings/projectMetrics/projectMetrics.tsx +++ b/static/app/views/settings/projectMetrics/projectMetrics.tsx @@ -42,7 +42,6 @@ import TextBlock from 'sentry/views/settings/components/text/textBlock'; import PermissionAlert from 'sentry/views/settings/project/permissionAlert'; import {useAccess} from 'sentry/views/settings/projectMetrics/access'; import {BlockButton} from 'sentry/views/settings/projectMetrics/blockButton'; -import {CardinalityLimit} from 'sentry/views/settings/projectMetrics/cardinalityLimit'; import { MetricsExtractionRuleEditModal, modalCss, @@ -150,8 +149,6 @@ function ProjectMetrics({project, location}: Props) { <PermissionAlert project={project} /> - <CardinalityLimit project={project} /> - {hasExtractionRules && ( <Fragment> <ExtractionRulesSearchWrapper>
fcf3e0dc838ec2ba986839ddd60ede1b2f2abbf0
2022-06-13 12:48:06
dependabot[bot]
chore(deps): bump @testing-library/jest-dom from 5.16.1 to 5.16.4 (#35402)
false
bump @testing-library/jest-dom from 5.16.1 to 5.16.4 (#35402)
chore
diff --git a/package.json b/package.json index 1b7f66673163e3..1e8dde58844946 100644 --- a/package.json +++ b/package.json @@ -30,7 +30,7 @@ "@sentry/rrweb": "^0.3.3", "@sentry/tracing": "7.0.0", "@sentry/utils": "7.0.0", - "@testing-library/jest-dom": "^5.16.1", + "@testing-library/jest-dom": "^5.16.4", "@testing-library/react": "^12.1.2", "@testing-library/react-hooks": "^8.0.0", "@testing-library/user-event": "^13.5.0", diff --git a/yarn.lock b/yarn.lock index 5f15c164f366a3..963b63f24c8d2f 100644 --- a/yarn.lock +++ b/yarn.lock @@ -3948,10 +3948,10 @@ lz-string "^1.4.4" pretty-format "^27.0.2" -"@testing-library/jest-dom@^5.16.1": - version "5.16.1" - resolved "https://registry.yarnpkg.com/@testing-library/jest-dom/-/jest-dom-5.16.1.tgz#3db7df5ae97596264a7da9696fe14695ba02e51f" - integrity sha512-ajUJdfDIuTCadB79ukO+0l8O+QwN0LiSxDaYUTI4LndbbUsGi6rWU1SCexXzBA2NSjlVB9/vbkasQIL3tmPBjw== +"@testing-library/jest-dom@^5.16.4": + version "5.16.4" + resolved "https://registry.yarnpkg.com/@testing-library/jest-dom/-/jest-dom-5.16.4.tgz#938302d7b8b483963a3ae821f1c0808f872245cd" + integrity sha512-Gy+IoFutbMQcky0k+bqqumXZ1cTGswLsFqmNLzNdSKkU9KGV2u9oXhukCbbJ9/LRPKiqwxEE8VpV/+YZlfkPUA== dependencies: "@babel/runtime" "^7.9.2" "@types/testing-library__jest-dom" "^5.9.1"
1d4ac511f74fb279dc1d1e8d0923ac44bb3ecf60
2020-02-03 14:50:28
Markus Unterwaditzer
feat: Better validation for PII config (#16719)
false
Better validation for PII config (#16719)
feat
diff --git a/src/sentry/api/endpoints/project_details.py b/src/sentry/api/endpoints/project_details.py index eb01162dbfa354..10dd52a76c9a97 100644 --- a/src/sentry/api/endpoints/project_details.py +++ b/src/sentry/api/endpoints/project_details.py @@ -10,6 +10,7 @@ from django.utils import timezone from rest_framework import serializers, status from rest_framework.response import Response +from sentry_relay import validate_pii_config from sentry import features from sentry.utils.data_filters import FilterTypes @@ -175,8 +176,6 @@ def validate_relayPiiConfig(self, value): if not value: return value - from sentry import features - organization = self.context["project"].organization request = self.context["request"] has_datascrubbers_v2 = features.has( @@ -186,6 +185,12 @@ def validate_relayPiiConfig(self, value): raise serializers.ValidationError( "Organization does not have the datascrubbers-v2 feature enabled" ) + + try: + validate_pii_config(value) + except ValueError as e: + raise serializers.ValidationError(e) + return value def validate_builtinSymbolSources(self, value): diff --git a/src/sentry/static/sentry/app/data/forms/projectGeneralSettings.jsx b/src/sentry/static/sentry/app/data/forms/projectGeneralSettings.jsx index 0b3bb1aad492f8..ac473745842b82 100644 --- a/src/sentry/static/sentry/app/data/forms/projectGeneralSettings.jsx +++ b/src/sentry/static/sentry/app/data/forms/projectGeneralSettings.jsx @@ -359,6 +359,9 @@ export const fields = { ), visible: ({features}) => features.has('datascrubbers-v2'), validate: ({id, form}) => { + if (form[id] === '') { + return []; + } try { JSON.parse(form[id]); } catch (e) {
4e660d535274542cc830da2d629d5ac90b80b0c7
2020-01-29 20:59:39
Alberto Leal
fix(discover2): Fixes for Discover2 tables with 1 column (#16674)
false
Fixes for Discover2 tables with 1 column (#16674)
fix
diff --git a/src/sentry/static/sentry/app/components/gridEditable/gridHeadCell.tsx b/src/sentry/static/sentry/app/components/gridEditable/gridHeadCell.tsx index 4cc84f0250b503..e2fd27c6b7337a 100644 --- a/src/sentry/static/sentry/app/components/gridEditable/gridHeadCell.tsx +++ b/src/sentry/static/sentry/app/components/gridEditable/gridHeadCell.tsx @@ -17,6 +17,7 @@ export type GridHeadCellProps<Column> = { isLast: boolean; isEditing: boolean; + isDeletable: boolean; indexColumnOrder: number; column: Column; @@ -49,6 +50,7 @@ class GridHeadCell<Column extends GridColumnHeader> extends React.Component< > { static defaultProps = { isEditing: false, + isDeletable: true, }; state = { @@ -79,12 +81,18 @@ class GridHeadCell<Column extends GridColumnHeader> extends React.Component< renderButtonHoverDraggable() { const {isHovering} = this.state; - const {isEditing, isColumnDragging} = this.props; + const {isEditing, isDeletable, isColumnDragging} = this.props; if (!isEditing || !isHovering || isColumnDragging) { return null; } + const deleteButton = isDeletable ? ( + <GridHeadCellButtonHoverButton onClick={this.deleteColumn}> + <InlineSvg src="icon-trash" /> + </GridHeadCellButtonHoverButton> + ) : null; + return ( <React.Fragment> {/* Ensure that background is always at the top. The background must be @@ -100,9 +108,7 @@ class GridHeadCell<Column extends GridColumnHeader> extends React.Component< <GridHeadCellButtonHoverButton onClick={this.toggleModal}> <InlineSvg src="icon-edit-pencil" /> </GridHeadCellButtonHoverButton> - <GridHeadCellButtonHoverButton onClick={this.deleteColumn}> - <InlineSvg src="icon-trash" /> - </GridHeadCellButtonHoverButton> + {deleteButton} </div> <GridHeadCellButtonHoverDraggable diff --git a/src/sentry/static/sentry/app/components/gridEditable/index.tsx b/src/sentry/static/sentry/app/components/gridEditable/index.tsx index 72b580172cf05f..3a3e56600ad3ad 100644 --- a/src/sentry/static/sentry/app/components/gridEditable/index.tsx +++ b/src/sentry/static/sentry/app/components/gridEditable/index.tsx @@ -444,7 +444,6 @@ class GridEditable< // Ensure that the last column cannot be removed const numColumn = columnOrder.length; - const enableEdit = isEditing && numColumn > 1; const prependColumns = grid.renderPrependColumns ? grid.renderPrependColumns(true) @@ -463,7 +462,8 @@ class GridEditable< isLast={columnOrder.length - 1 === i} key={`${i}.${column.key}`} isColumnDragging={this.props.isColumnDragging} - isEditing={enableEdit} + isEditing={isEditing} + isDeletable={numColumn > 1} indexColumnOrder={i} column={column} gridHeadCellButtonProps={this.props.gridHeadCellButtonProps || {}}
215a0fde5a07732253a383c37be8f460b6161ca3
2021-04-30 17:43:27
Markus Unterwaditzer
ref: Basic typing for grouping internals (#25690)
false
Basic typing for grouping internals (#25690)
ref
diff --git a/mypy.ini b/mypy.ini index c1913f9f8e4b01..1f65f21f13551b 100644 --- a/mypy.ini +++ b/mypy.ini @@ -1,12 +1,12 @@ [mypy] python_version = 3.6 files = src/sentry/api/bases/external_actor.py, - src/sentry/api/endpoints/project_codeowners.py, src/sentry/api/endpoints/external_team.py, src/sentry/api/endpoints/external_team_details.py, src/sentry/api/endpoints/external_user.py, src/sentry/api/endpoints/external_user_details.py, src/sentry/api/endpoints/organization_events_trace.py, + src/sentry/api/endpoints/project_codeowners.py, src/sentry/api/serializers/base.py, src/sentry/api/serializers/models/external_actor.py, src/sentry/api/serializers/models/integration.py, @@ -15,6 +15,8 @@ files = src/sentry/api/bases/external_actor.py, src/sentry/api/serializers/models/team.py, src/sentry/api/validators/external_actor.py, src/sentry/api/validators/notifications.py, + src/sentry/grouping/strategies/base.py, + src/sentry/grouping/strategies/message.py, src/sentry/notifications/**/*.py, src/sentry/snuba/outcomes.py, src/sentry/snuba/query_subscription_consumer.py, diff --git a/src/sentry/grouping/strategies/base.py b/src/sentry/grouping/strategies/base.py index 95709e98814e12..569069a077f516 100644 --- a/src/sentry/grouping/strategies/base.py +++ b/src/sentry/grouping/strategies/base.py @@ -1,75 +1,100 @@ import inspect +from typing import Any, Callable, Dict, Iterator, List, Optional, Sequence, Type, Union import sentry_sdk from sentry import projectoptions +from sentry.eventstore.models import Event +from sentry.grouping.component import GroupingComponent from sentry.grouping.enhancer import Enhancements +from sentry.interfaces.base import Interface -STRATEGIES = {} - +STRATEGIES: Dict[str, "Strategy"] = {} RISK_LEVEL_LOW = 0 RISK_LEVEL_MEDIUM = 1 RISK_LEVEL_HIGH = 2 +Risk = int # TODO: make enum or union of literals + +# XXX: Want to make ContextDict typeddict but also want to type/overload dict +# API on GroupingContext +ContextValue = Any +ContextDict = Dict[str, ContextValue] + DEFAULT_GROUPING_ENHANCEMENTS_BASE = "common:2019-03-23" +ReturnedVariants = Dict[str, GroupingComponent] +StrategyFunc = Callable[..., ReturnedVariants] # TODO -def strategy(id=None, ids=None, interfaces=None, name=None, score=None): + +def strategy( + id: Optional[str] = None, + ids: Optional[Sequence[str]] = None, + interfaces: Optional[Sequence[str]] = None, + score: Optional[int] = None, +) -> Callable[[StrategyFunc], "Strategy"]: """Registers a strategy""" - if interfaces is None: + + if not interfaces: raise TypeError("interfaces is required") - if name is None: - if len(interfaces) != 1: - raise RuntimeError("%r requires a name" % id) - name = interfaces[0] + name = interfaces[0] if id is not None: if ids is not None: raise TypeError("id and ids given") ids = [id] - def decorator(f): + if not ids: + raise TypeError("neither id nor ids given") + + def decorator(f: StrategyFunc) -> Strategy: + assert interfaces + assert ids + for id in ids: STRATEGIES[id] = rv = Strategy( id=id, name=name, interfaces=interfaces, score=score, func=f ) + return rv return decorator class GroupingContext: - def __init__(self, strategy_config): + def __init__(self, strategy_config: "StrategyConfiguration"): self._stack = [strategy_config.initial_context] self.config = strategy_config self.push() self["variant"] = None - def __setitem__(self, key, value): + def __setitem__(self, key: str, value: ContextValue) -> None: self._stack[-1][key] = value - def __getitem__(self, key): + def __getitem__(self, key: str) -> ContextValue: for d in reversed(self._stack): if key in d: return d[key] raise KeyError(key) - def __enter__(self): + def __enter__(self) -> "GroupingContext": self.push() return self - def __exit__(self, exc_type, exc_value, tb): + def __exit__(self, exc_type: Type[Exception], exc_value: Exception, tb: Any) -> None: self.pop() - def push(self): + def push(self) -> None: self._stack.append({}) - def pop(self): + def pop(self) -> None: self._stack.pop() - def get_grouping_component(self, interface, *args, **kwargs): + def get_grouping_component( + self, interface: Interface, *args: Any, **kwargs: Any + ) -> Union[GroupingComponent, ReturnedVariants]: """Invokes a delegate grouping strategy. If no such delegate is configured a fallback grouping component is returned. """ @@ -92,7 +117,7 @@ def get_grouping_component(self, interface, *args, **kwargs): return rv -def lookup_strategy(strategy_id): +def lookup_strategy(strategy_id: str) -> "Strategy": """Looks up a strategy by id.""" try: return STRATEGIES[strategy_id] @@ -100,48 +125,48 @@ def lookup_strategy(strategy_id): raise LookupError("Unknown strategy %r" % strategy_id) -def flatten_variants_from_component(component): - """Given a component extracts variants from it if that component is - a variant provider. Otherwise this just returns the root component. - """ - if not component.variant_provider: - return {component.id: component} - return {c.id: c for c in component.values} - - class Strategy: """Baseclass for all strategies.""" - def __init__(self, id, name, interfaces, score, func): + def __init__( + self, + id: str, + name: str, + interfaces: Sequence[str], + score: Optional[int], + func: StrategyFunc, + ): self.id = id self.strategy_class = id.split(":", 1)[0] self.name = name self.interfaces = interfaces self.score = score self.func = func - self.variant_processor_func = None + self.variant_processor_func: Optional[StrategyFunc] = None - def __repr__(self): + def __repr__(self) -> str: return f"<{self.__class__.__name__} id={self.id!r}>" - def _invoke(self, func, *args, **kwargs): + def _invoke(self, func: StrategyFunc, *args: Any, **kwargs: Any) -> ReturnedVariants: # We forcefully override strategy here. This lets a strategy # function always access its metadata and directly forward it to # subcomponents without having to filter out strategy. kwargs["strategy"] = self return func(*args, **kwargs) - def __call__(self, *args, **kwargs): + def __call__(self, *args: Any, **kwargs: Any) -> ReturnedVariants: return self._invoke(self.func, *args, **kwargs) - def variant_processor(self, func): + def variant_processor(self, func: StrategyFunc) -> StrategyFunc: """Registers a variant reducer function that can be used to postprocess all variants created from this strategy. """ self.variant_processor_func = func return func - def get_grouping_component(self, event, context, variant=None): + def get_grouping_component( + self, event: Event, context: GroupingContext, variant: Optional[str] = None + ) -> Optional[ReturnedVariants]: """Given a specific variant this calculates the grouping component.""" args = [] for iface_path in self.interfaces: @@ -155,7 +180,9 @@ def get_grouping_component(self, event, context, variant=None): context["variant"] = variant return self(event=event, context=context, *args) - def get_grouping_component_variants(self, event, context): + def get_grouping_component_variants( + self, event: Event, context: GroupingContext + ) -> ReturnedVariants: """This returns a dictionary of all components by variant that this strategy can produce. """ @@ -223,59 +250,59 @@ def get_grouping_component_variants(self, event, context): class StrategyConfiguration: - id = None - base = None + id: Optional[str] = None + base: Optional[Type["StrategyConfiguration"]] = None config_class = None - strategies = {} - delegates = {} - changelog = None + strategies: Dict[str, Strategy] = {} + delegates: Dict[str, Strategy] = {} + changelog: Optional[str] = None hidden = False risk = RISK_LEVEL_LOW - initial_context = {} - enhancements_base: str = DEFAULT_GROUPING_ENHANCEMENTS_BASE + initial_context: ContextDict = {} + enhancements_base: Optional[str] = DEFAULT_GROUPING_ENHANCEMENTS_BASE - def __init__(self, enhancements=None, **extra): + def __init__(self, enhancements: Optional[str] = None, **extra: Any): if enhancements is None: - enhancements = Enhancements([]) + enhancements_instance = Enhancements([]) else: - enhancements = Enhancements.loads(enhancements) - self.enhancements = enhancements + enhancements_instance = Enhancements.loads(enhancements) + self.enhancements = enhancements_instance - def __repr__(self): + def __repr__(self) -> str: return f"<{self.__class__.__name__} {self.id!r}>" - def iter_strategies(self): + def iter_strategies(self) -> Iterator[Strategy]: """Iterates over all strategies by highest score to lowest.""" - return iter(sorted(self.strategies.values(), key=lambda x: -x.score)) + return iter(sorted(self.strategies.values(), key=lambda x: x.score and -x.score or 0)) @classmethod - def as_dict(self): + def as_dict(cls) -> Dict[str, Any]: return { - "id": self.id, - "base": self.base.id if self.base else None, - "strategies": sorted(self.strategies), - "changelog": self.changelog, - "delegates": sorted(x.id for x in self.delegates.values()), - "hidden": self.hidden, - "risk": self.risk, + "id": cls.id, + "base": cls.base.id if cls.base else None, + "strategies": sorted(cls.strategies), + "changelog": cls.changelog, + "delegates": sorted(x.id for x in cls.delegates.values()), + "hidden": cls.hidden, + "risk": cls.risk, "latest": projectoptions.lookup_well_known_key("sentry:grouping_config").get_default( epoch=projectoptions.LATEST_EPOCH ) - == self.id, + == cls.id, } def create_strategy_configuration( - id, - strategies=None, - delegates=None, - changelog=None, - hidden=False, - base=None, - risk=None, - initial_context=None, - enhancements_base=None, -): + id: str, + strategies: Optional[Sequence[str]] = None, + delegates: Optional[Sequence[str]] = None, + changelog: Optional[str] = None, + hidden: bool = False, + base: Optional[Type[StrategyConfiguration]] = None, + risk: Optional[Risk] = None, + initial_context: Optional[ContextDict] = None, + enhancements_base: Optional[str] = None, +) -> Type[StrategyConfiguration]: """Declares a new strategy configuration. Values can be inherited from a base configuration. For strategies if there is @@ -300,7 +327,7 @@ class NewStrategyConfiguration(StrategyConfiguration): NewStrategyConfiguration.risk = risk NewStrategyConfiguration.hidden = hidden - by_class = {} + by_class: Dict[str, List[str]] = {} for strategy in NewStrategyConfiguration.strategies.values(): by_class.setdefault(strategy.strategy_class, []).append(strategy.id) @@ -335,7 +362,7 @@ class NewStrategyConfiguration(StrategyConfiguration): return NewStrategyConfiguration -def produces_variants(variants): +def produces_variants(variants: Sequence[str]) -> Callable[[StrategyFunc], StrategyFunc]: """ A grouping strategy can either: @@ -365,8 +392,8 @@ def produces_variants(variants): @produces_variants(["!system", "app"]) """ - def decorator(f): - def inner(*args, **kwargs): + def decorator(f: StrategyFunc) -> StrategyFunc: + def inner(*args: Any, **kwargs: Any) -> ReturnedVariants: return call_with_variants(f, variants, *args, **kwargs) return inner @@ -374,7 +401,9 @@ def inner(*args, **kwargs): return decorator -def call_with_variants(f, variants, *args, **kwargs): +def call_with_variants( + f: StrategyFunc, variants: Sequence[str], *args: Any, **kwargs: Any +) -> ReturnedVariants: context = kwargs["context"] if context["variant"] is not None: # For the case where the variant is already determined, we act as a @@ -390,9 +419,9 @@ def call_with_variants(f, variants, *args, **kwargs): for variant in variants: with context: context["variant"] = variant.lstrip("!") - variants = f(*args, **kwargs) - assert len(variants) == 1 - component = variants[variant.lstrip("!")] + rv_variants = f(*args, **kwargs) + assert len(rv_variants) == 1 + component = rv_variants[variant.lstrip("!")] if component is None: continue diff --git a/src/sentry/grouping/strategies/message.py b/src/sentry/grouping/strategies/message.py index 4f911f2b0fe4ba..51f79305f93505 100644 --- a/src/sentry/grouping/strategies/message.py +++ b/src/sentry/grouping/strategies/message.py @@ -1,9 +1,16 @@ import re from itertools import islice +from typing import Any, Match from sentry.grouping.component import GroupingComponent -from sentry.grouping.strategies.base import produces_variants, strategy +from sentry.grouping.strategies.base import ( + GroupingContext, + ReturnedVariants, + produces_variants, + strategy, +) from sentry.grouping.strategies.similarity_encoders import text_shingle_encoder +from sentry.interfaces.message import Message _irrelevant_re = re.compile( r"""(?x) @@ -90,12 +97,12 @@ ) -def trim_message_for_grouping(string): +def trim_message_for_grouping(string: str) -> str: s = "\n".join(islice((x for x in string.splitlines() if x.strip()), 2)).strip() if s != string: s += "..." - def _handle_match(match): + def _handle_match(match: Match[str]) -> str: for key, value in match.groupdict().items(): if value is not None: return "<%s>" % key @@ -106,7 +113,9 @@ def _handle_match(match): @strategy(id="message:v1", interfaces=["message"], score=0) @produces_variants(["default"]) -def message_v1(message_interface, context, **meta): +def message_v1( + message_interface: Message, context: GroupingContext, **meta: Any +) -> ReturnedVariants: if context["trim_message"]: message_in = message_interface.message or message_interface.formatted or "" message_trimmed = trim_message_for_grouping(message_in)
36fcc89dd7c558021c41b951cb519730ee69f0c5
2024-05-04 02:09:10
Yagiz Nizipli
perf: create orjson experiment for relay-py (#69613)
false
create orjson experiment for relay-py (#69613)
perf
diff --git a/src/sentry/api/endpoints/data_scrubbing_selector_suggestions.py b/src/sentry/api/endpoints/data_scrubbing_selector_suggestions.py index b5e43e04ef256a..aa09d70baedcc8 100644 --- a/src/sentry/api/endpoints/data_scrubbing_selector_suggestions.py +++ b/src/sentry/api/endpoints/data_scrubbing_selector_suggestions.py @@ -10,6 +10,7 @@ from sentry.api.base import region_silo_endpoint from sentry.api.bases.organization import OrganizationEndpoint from sentry.eventstore.models import Event +from sentry.utils.json import methods_for_experiment @region_silo_endpoint @@ -49,9 +50,12 @@ def get(self, request: Request, organization) -> Response: node_ids = [Event.generate_node_id(p, event_id) for p in project_ids] all_data = nodestore.backend.get_multi(node_ids) + json_loads, json_dumps = methods_for_experiment("relay.enable-orjson") data: dict[str, Any] for data in filter(None, all_data.values()): - for selector in pii_selector_suggestions_from_event(data): + for selector in pii_selector_suggestions_from_event( + data, json_loads=json_loads, json_dumps=json_dumps + ): examples_ = suggestions.setdefault(selector["path"], []) if selector["value"]: examples_.append(selector["value"]) diff --git a/src/sentry/datascrubbing.py b/src/sentry/datascrubbing.py index fcbe8ac1615504..4d931d91116c5d 100644 --- a/src/sentry/datascrubbing.py +++ b/src/sentry/datascrubbing.py @@ -78,7 +78,10 @@ def get_all_pii_configs(project): if pii_config: yield pii_config - yield convert_datascrubbing_config(get_datascrubbing_settings(project)) + settings = get_datascrubbing_settings(project) + + json_loads, json_dumps = json.methods_for_experiment("relay.enable-orjson") + yield convert_datascrubbing_config(settings, json_dumps=json_dumps, json_loads=json_loads) @sentry_sdk.tracing.trace @@ -95,7 +98,8 @@ def scrub_data(project, event): metrics.distribution("datascrubbing.config.rules.size", total_rules) - event = pii_strip_event(config, event) + json_loads, json_dumps = json.methods_for_experiment("relay.enable-orjson") + event = pii_strip_event(config, event, json_loads=json_loads, json_dumps=json_dumps) return event diff --git a/src/sentry/event_manager.py b/src/sentry/event_manager.py index 9e575cee5c5de0..92874f4672ef27 100644 --- a/src/sentry/event_manager.py +++ b/src/sentry/event_manager.py @@ -375,6 +375,8 @@ def _normalize_impl(self, project_id: int | None = None) -> None: from sentry_relay.processing import StoreNormalizer + json_loads, json_dumps = json.methods_for_experiment("relay.enable-orjson") + rust_normalizer = StoreNormalizer( project_id=self._project.id if self._project else project_id, client_ip=self._client_ip, @@ -386,11 +388,14 @@ def _normalize_impl(self, project_id: int | None = None) -> None: remove_other=self._remove_other, normalize_user_agent=True, sent_at=self.sent_at.isoformat() if self.sent_at is not None else None, + json_dumps=json_dumps, **DEFAULT_STORE_NORMALIZER_ARGS, ) pre_normalize_type = self._data.get("type") - self._data = CanonicalKeyDict(rust_normalizer.normalize_event(dict(self._data))) + self._data = CanonicalKeyDict( + rust_normalizer.normalize_event(dict(self._data), json_loads=json_loads) + ) # XXX: This is a hack to make generic events work (for now?). I'm not sure whether we should # include this in the rust normalizer, since we don't want people sending us these via the # sdk. diff --git a/src/sentry/integrations/slack/message_builder/issues.py b/src/sentry/integrations/slack/message_builder/issues.py index 62361414b493de..0fb89edced10f4 100644 --- a/src/sentry/integrations/slack/message_builder/issues.py +++ b/src/sentry/integrations/slack/message_builder/issues.py @@ -173,7 +173,8 @@ def format_release_tag(value: str, event: GroupEvent | Group): """Format the release tag using the short version and make it a link""" path = f"/releases/{value}/" url = event.project.organization.absolute_url(path) - release_description = parse_release(value).get("description") + json_loads, _ = json.methods_for_experiment("relay.enable-orjson") + release_description = parse_release(value, json_loads=json_loads).get("description") return f"<{url}|{release_description}>" diff --git a/src/sentry/integrations/slack/message_builder/notifications/daily_summary.py b/src/sentry/integrations/slack/message_builder/notifications/daily_summary.py index c2aafcd6580c79..c692fcb4b51baf 100644 --- a/src/sentry/integrations/slack/message_builder/notifications/daily_summary.py +++ b/src/sentry/integrations/slack/message_builder/notifications/daily_summary.py @@ -52,7 +52,10 @@ def linkify_error_title(self, group): def linkify_release(self, release, organization): path = f"/releases/{release.version}/" url = organization.absolute_url(path) - release_description = parse_release(release.version).get("description") + json_loads, _ = json.methods_for_experiment("relay.enable-orjson") + release_description = parse_release(release.version, json_loads=json_loads).get( + "description" + ) return f":rocket: *<{url}|Release {release_description}>*\n" def truncate_text(self, text): diff --git a/src/sentry/models/event.py b/src/sentry/models/event.py index 01fbc4d9a7efc4..49bf256b25e219 100644 --- a/src/sentry/models/event.py +++ b/src/sentry/models/event.py @@ -1,6 +1,7 @@ from sentry_relay.processing import StoreNormalizer from sentry.db.models import NodeData +from sentry.utils import json from sentry.utils.canonical import CanonicalKeyDict @@ -20,10 +21,13 @@ def __init__(self, data, skip_renormalization=False, **kwargs): ) if not skip_renormalization and not is_renormalized: - normalizer = StoreNormalizer(is_renormalize=True, enable_trimming=False) data = dict(data) pre_normalize_type = data.get("type") - data = normalizer.normalize_event(data) + json_loads, json_dumps = json.methods_for_experiment("relay.enable-orjson") + normalizer = StoreNormalizer( + is_renormalize=True, enable_trimming=False, json_dumps=json_dumps + ) + data = normalizer.normalize_event(data, json_loads=json_loads) # XXX: This is a hack to make generic events work (for now?). I'm not sure whether we # should include this in the rust normalizer, since we don't want people sending us # these via the sdk. diff --git a/src/sentry/models/groupresolution.py b/src/sentry/models/groupresolution.py index 3233386340913f..1eed58dcde129e 100644 --- a/src/sentry/models/groupresolution.py +++ b/src/sentry/models/groupresolution.py @@ -15,7 +15,7 @@ ) from sentry.models.release import Release, follows_semver_versioning_scheme from sentry.models.releases.constants import DB_VERSION_LENGTH -from sentry.utils import metrics +from sentry.utils import json, metrics @region_silo_model @@ -116,8 +116,13 @@ def compare_release_dates_for_in_next_release(res_release, res_release_datetime, # If current_release_version == release.version => 0 # If current_release_version < release.version => -1 # If current_release_version > release.version => 1 - current_release_raw = parse_release(current_release_version).get("version_raw") - release_raw = parse_release(release.version).get("version_raw") + json_loads, _ = json.methods_for_experiment("relay.enable-orjson") + current_release_raw = parse_release( + current_release_version, json_loads=json_loads + ).get("version_raw") + release_raw = parse_release(release.version, json_loads=json_loads).get( + "version_raw" + ) return compare_version_relay(current_release_raw, release_raw) >= 0 except RelayError: ... @@ -157,8 +162,13 @@ def compare_release_dates_for_in_next_release(res_release, res_release_datetime, try: # A resolution only exists if the resolved release is greater (in semver # terms) than the provided release - res_release_raw = parse_release(res_release_version).get("version_raw") - release_raw = parse_release(release.version).get("version_raw") + json_loads, _ = json.methods_for_experiment("relay.enable-orjson") + res_release_raw = parse_release(res_release_version, json_loads=json_loads).get( + "version_raw" + ) + release_raw = parse_release(release.version, json_loads=json_loads).get( + "version_raw" + ) return compare_version_relay(res_release_raw, release_raw) == 1 except RelayError: ... diff --git a/src/sentry/models/release.py b/src/sentry/models/release.py index 3c3933a3b4de96..ee2ef63a5e6ce4 100644 --- a/src/sentry/models/release.py +++ b/src/sentry/models/release.py @@ -48,7 +48,7 @@ from sentry.models.releases.util import ReleaseQuerySet, SemverFilter, SemverVersion from sentry.services.hybrid_cloud.user import RpcUser from sentry.signals import issue_resolved -from sentry.utils import metrics +from sentry.utils import json, metrics from sentry.utils.cache import cache from sentry.utils.db import atomic_transaction from sentry.utils.hashlib import hash_values, md5_text @@ -346,7 +346,8 @@ def is_semver_version(version): return False try: - version_info = parse_release(version) + json_loads, _ = json.methods_for_experiment("relay.enable-orjson") + version_info = parse_release(version, json_loads=json_loads) version_parsed = version_info.get("version_parsed") return version_parsed is not None and all( validate_bigint(version_parsed[field]) @@ -488,7 +489,8 @@ def _get_or_create_impl(cls, project, version, date_added, metric_tags): @cached_property def version_info(self): try: - return parse_release(self.version) + json_loads, _ = json.methods_for_experiment("relay.enable-orjson") + return parse_release(self.version, json_loads=json_loads) except RelayError: # This can happen on invalid legacy releases return None diff --git a/src/sentry/notifications/notifications/activity/regression.py b/src/sentry/notifications/notifications/activity/regression.py index 3055e1dd7fe7d0..9a1e35c69f9724 100644 --- a/src/sentry/notifications/notifications/activity/regression.py +++ b/src/sentry/notifications/notifications/activity/regression.py @@ -8,6 +8,7 @@ from sentry.models.activity import Activity from sentry.types.integrations import ExternalProviders +from sentry.utils.json import methods_for_experiment from .base import GroupActivityNotification @@ -19,7 +20,8 @@ class RegressionActivityNotification(GroupActivityNotification): def __init__(self, activity: Activity) -> None: super().__init__(activity) self.version = self.activity.data.get("version", "") - self.version_parsed = parse_release(self.version)["description"] + json_loads, _ = methods_for_experiment("relay.enable-orjson") + self.version_parsed = parse_release(self.version, json_loads=json_loads)["description"] def get_description(self) -> tuple[str, str | None, Mapping[str, Any]]: text_message, html_message, params = "{author} marked {an issue} as a regression", None, {} diff --git a/src/sentry/notifications/notifications/activity/release.py b/src/sentry/notifications/notifications/activity/release.py index 07ec1b87bd10eb..f89a770b0fed14 100644 --- a/src/sentry/notifications/notifications/activity/release.py +++ b/src/sentry/notifications/notifications/activity/release.py @@ -24,6 +24,7 @@ from sentry.services.hybrid_cloud.actor import ActorType, RpcActor from sentry.services.hybrid_cloud.user.service import user_service from sentry.types.integrations import ExternalProviders +from sentry.utils.json import methods_for_experiment from .base import ActivityNotification @@ -63,7 +64,8 @@ def __init__(self, activity: Activity) -> None: self.group_counts_by_project = get_group_counts_by_project(self.release, self.projects) self.version = self.release.version - self.version_parsed = parse_release(self.version)["description"] + json_loads, _ = methods_for_experiment("relay.enable-orjson") + self.version_parsed = parse_release(self.version, json_loads=json_loads)["description"] def get_participants_with_group_subscription_reason(self) -> ParticipantMap: return get_participants_for_release(self.projects, self.organization, self.user_ids) diff --git a/src/sentry/notifications/notifications/activity/resolved_in_release.py b/src/sentry/notifications/notifications/activity/resolved_in_release.py index c394b19affc559..2d69e4532de78c 100644 --- a/src/sentry/notifications/notifications/activity/resolved_in_release.py +++ b/src/sentry/notifications/notifications/activity/resolved_in_release.py @@ -8,6 +8,7 @@ from sentry.models.activity import Activity from sentry.types.integrations import ExternalProviders +from sentry.utils.json import methods_for_experiment from .base import GroupActivityNotification @@ -19,7 +20,8 @@ class ResolvedInReleaseActivityNotification(GroupActivityNotification): def __init__(self, activity: Activity) -> None: super().__init__(activity) self.version = self.activity.data.get("version", "") - self.version_parsed = parse_release(self.version)["description"] + json_loads, _ = methods_for_experiment("relay.enable-orjson") + self.version_parsed = parse_release(self.version, json_loads=json_loads)["description"] def get_description(self) -> tuple[str, str | None, Mapping[str, Any]]: if self.version: diff --git a/src/sentry/options/defaults.py b/src/sentry/options/defaults.py index 4d0a5353a2b8aa..d54d3588623edf 100644 --- a/src/sentry/options/defaults.py +++ b/src/sentry/options/defaults.py @@ -650,6 +650,7 @@ register("event-manager.enable-orjson", default=0.0, flags=FLAG_AUTOMATOR_MODIFIABLE) register("eventstore.enable-orjson", default=0.0, flags=FLAG_AUTOMATOR_MODIFIABLE) register("flagpole.enable-orjson", default=0.0, flags=FLAG_AUTOMATOR_MODIFIABLE) +register("relay.enable-orjson", default=0.0, flags=FLAG_AUTOMATOR_MODIFIABLE) # Kafka Publisher register("kafka-publisher.raw-event-sample-rate", default=0.0, flags=FLAG_AUTOMATOR_MODIFIABLE) diff --git a/src/sentry/search/events/filter.py b/src/sentry/search/events/filter.py index 8696686fe3c3b3..912af045a1f5d8 100644 --- a/src/sentry/search/events/filter.py +++ b/src/sentry/search/events/filter.py @@ -48,6 +48,7 @@ ) from sentry.search.events.fields import FIELD_ALIASES, FUNCTIONS, resolve_field from sentry.search.utils import parse_release +from sentry.utils.json import methods_for_experiment from sentry.utils.snuba import FUNCTION_TO_OPERATOR, OPERATOR_TO_FUNCTION, SNUBA_AND, SNUBA_OR from sentry.utils.strings import oxfordize_list from sentry.utils.validators import INVALID_ID_DETAILS, INVALID_SPAN_ID, WILDCARD_NOT_ALLOWED @@ -523,7 +524,8 @@ def parse_semver(version, operator) -> SemverFilter: raise InvalidSearchQuery("Invalid operation 'IN' for semantic version filter.") version = version if "@" in version else f"{SEMVER_FAKE_PACKAGE}@{version}" - parsed = parse_release_relay(version) + json_loads, _ = methods_for_experiment("relay.enable-orjson") + parsed = parse_release_relay(version, json_loads=json_loads) parsed_version = parsed.get("version_parsed") if parsed_version: # Convert `pre` to always be a string diff --git a/src/sentry/tasks/store.py b/src/sentry/tasks/store.py index 4df36d7454d106..980d1c3229c7fc 100644 --- a/src/sentry/tasks/store.py +++ b/src/sentry/tasks/store.py @@ -27,7 +27,7 @@ from sentry.stacktraces.processing import process_stacktraces, should_process_for_stacktraces from sentry.tasks.base import instrumented_task from sentry.types.activity import ActivityType -from sentry.utils import metrics +from sentry.utils import json, metrics from sentry.utils.canonical import CANONICAL_TYPES, CanonicalKeyDict from sentry.utils.dates import to_datetime from sentry.utils.safe import safe_execute @@ -307,10 +307,14 @@ def is_process_disabled(project_id: int, event_id: str, platform: str) -> bool: @sentry_sdk.tracing.trace def normalize_event(data: Any) -> Any: + json_loads, json_dumps = json.methods_for_experiment("relay.enable-orjson") normalizer = StoreNormalizer( - remove_other=False, is_renormalize=True, **DEFAULT_STORE_NORMALIZER_ARGS + remove_other=False, + is_renormalize=True, + json_dumps=json_dumps, + **DEFAULT_STORE_NORMALIZER_ARGS, ) - return normalizer.normalize_event(dict(data)) + return normalizer.normalize_event(dict(data), json_loads=json_loads) def do_process_event( diff --git a/src/sentry/utils/json.py b/src/sentry/utils/json.py index dd06c380a7dcb2..0b1a1723b8a41b 100644 --- a/src/sentry/utils/json.py +++ b/src/sentry/utils/json.py @@ -5,8 +5,9 @@ import contextlib import datetime import decimal +import json # noqa: S003 import uuid -from collections.abc import Generator, Mapping +from collections.abc import Callable, Generator, Mapping from enum import Enum from typing import IO, TYPE_CHECKING, Any, NoReturn, TypeVar, overload @@ -174,6 +175,18 @@ def dumps_htmlsafe(value: object) -> SafeString: return mark_safe(_default_escaped_encoder.encode(value)) +# TODO: remove this when orjson experiment is successful +def methods_for_experiment( + option_name: str, +) -> tuple[Callable[[str | bytes], Any], Callable[[Any], Any]]: + from sentry.features.rollout import in_random_rollout + + if in_random_rollout(option_name): + return orjson.loads, orjson.dumps + else: + return json.loads, json.dumps + + @overload def prune_empty_keys(obj: None) -> None: ... @@ -211,4 +224,7 @@ def prune_empty_keys(obj: Mapping[TKey, TValue | None] | None) -> dict[TKey, TVa "load", "loads", "prune_empty_keys", + "methods_for_experiment", + "loads_experimental", + "dumps_experimental", ) diff --git a/tests/sentry/event_manager/interfaces/test_contexts.py b/tests/sentry/event_manager/interfaces/test_contexts.py index 2fc78f4213ef68..b6296e51006034 100644 --- a/tests/sentry/event_manager/interfaces/test_contexts.py +++ b/tests/sentry/event_manager/interfaces/test_contexts.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_csp.py b/tests/sentry/event_manager/interfaces/test_csp.py index b7e5d72a2cdcbc..851a69ab577579 100644 --- a/tests/sentry/event_manager/interfaces/test_csp.py +++ b/tests/sentry/event_manager/interfaces/test_csp.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager, get_event_type, materialize_metadata +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_debug_meta.py b/tests/sentry/event_manager/interfaces/test_debug_meta.py index e77d7089e6ef27..66e3af07749eec 100644 --- a/tests/sentry/event_manager/interfaces/test_debug_meta.py +++ b/tests/sentry/event_manager/interfaces/test_debug_meta.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_exception.py b/tests/sentry/event_manager/interfaces/test_exception.py index 1b30c09328f088..3745203acefb9e 100644 --- a/tests/sentry/event_manager/interfaces/test_exception.py +++ b/tests/sentry/event_manager/interfaces/test_exception.py @@ -4,6 +4,13 @@ from sentry.event_manager import EventManager from sentry.interfaces.exception import Exception from sentry.stacktraces.processing import normalize_stacktraces_for_grouping +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_expectstaple.py b/tests/sentry/event_manager/interfaces/test_expectstaple.py index 22220e78e52714..45663984a99d43 100644 --- a/tests/sentry/event_manager/interfaces/test_expectstaple.py +++ b/tests/sentry/event_manager/interfaces/test_expectstaple.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager, get_event_type, materialize_metadata +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_geo.py b/tests/sentry/event_manager/interfaces/test_geo.py index 5a318f840543c2..58b244fd9ba6ec 100644 --- a/tests/sentry/event_manager/interfaces/test_geo.py +++ b/tests/sentry/event_manager/interfaces/test_geo.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_mechanism.py b/tests/sentry/event_manager/interfaces/test_mechanism.py index 20998e79c5bf32..4a1ce9bbd366c8 100644 --- a/tests/sentry/event_manager/interfaces/test_mechanism.py +++ b/tests/sentry/event_manager/interfaces/test_mechanism.py @@ -3,6 +3,13 @@ from sentry import eventstore from sentry.event_manager import EventManager from sentry.interfaces.exception import upgrade_legacy_mechanism +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_message.py b/tests/sentry/event_manager/interfaces/test_message.py index 2d8f9430fb999b..1e89a43a6bfebb 100644 --- a/tests/sentry/event_manager/interfaces/test_message.py +++ b/tests/sentry/event_manager/interfaces/test_message.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_sdk.py b/tests/sentry/event_manager/interfaces/test_sdk.py index 26ce0892bfe385..ff2304ca7693e4 100644 --- a/tests/sentry/event_manager/interfaces/test_sdk.py +++ b/tests/sentry/event_manager/interfaces/test_sdk.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_spans.py b/tests/sentry/event_manager/interfaces/test_spans.py index df341d661008d5..1ae1e76daee8be 100644 --- a/tests/sentry/event_manager/interfaces/test_spans.py +++ b/tests/sentry/event_manager/interfaces/test_spans.py @@ -2,11 +2,18 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options START_TIME = 1562873192.624 END_TIME = 1562873194.624 [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield + + @pytest.fixture def make_spans_snapshot(insta_snapshot): def inner(data): diff --git a/tests/sentry/event_manager/interfaces/test_stacktrace.py b/tests/sentry/event_manager/interfaces/test_stacktrace.py index b1b375fc822d55..63d389f25584d5 100644 --- a/tests/sentry/event_manager/interfaces/test_stacktrace.py +++ b/tests/sentry/event_manager/interfaces/test_stacktrace.py @@ -5,6 +5,13 @@ from sentry import eventstore from sentry.event_manager import EventManager from sentry.interfaces.stacktrace import get_context, is_url +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield def test_is_url(): diff --git a/tests/sentry/event_manager/interfaces/test_template.py b/tests/sentry/event_manager/interfaces/test_template.py index 7cfcc8eca5fb8b..b87f2cca3cf148 100644 --- a/tests/sentry/event_manager/interfaces/test_template.py +++ b/tests/sentry/event_manager/interfaces/test_template.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_threads.py b/tests/sentry/event_manager/interfaces/test_threads.py index 8371f5ab87c52b..26c09a709d949e 100644 --- a/tests/sentry/event_manager/interfaces/test_threads.py +++ b/tests/sentry/event_manager/interfaces/test_threads.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/interfaces/test_user.py b/tests/sentry/event_manager/interfaces/test_user.py index 6bfcd5d2d9a9dc..5e489c4ffd153d 100644 --- a/tests/sentry/event_manager/interfaces/test_user.py +++ b/tests/sentry/event_manager/interfaces/test_user.py @@ -2,6 +2,13 @@ from sentry import eventstore from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield @pytest.fixture diff --git a/tests/sentry/event_manager/test_normalization.py b/tests/sentry/event_manager/test_normalization.py index 8ab91e1cfb2182..5a711545366331 100644 --- a/tests/sentry/event_manager/test_normalization.py +++ b/tests/sentry/event_manager/test_normalization.py @@ -6,6 +6,13 @@ from sentry.constants import DEFAULT_LOGGER_NAME, MAX_CULPRIT_LENGTH from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield def make_event(**kwargs): diff --git a/tests/sentry/event_manager/test_validate_data.py b/tests/sentry/event_manager/test_validate_data.py index b794b93a68ea95..914105fccb1ca2 100644 --- a/tests/sentry/event_manager/test_validate_data.py +++ b/tests/sentry/event_manager/test_validate_data.py @@ -5,6 +5,13 @@ from sentry.constants import MAX_CULPRIT_LENGTH, MAX_VERSION_LENGTH from sentry.event_manager import EventManager +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield def validate_and_normalize(data): diff --git a/tests/sentry/search/events/test_filter.py b/tests/sentry/search/events/test_filter.py index 73a342d7401814..3ac3414c572a01 100644 --- a/tests/sentry/search/events/test_filter.py +++ b/tests/sentry/search/events/test_filter.py @@ -28,9 +28,16 @@ from sentry.search.events.types import ParamsType, QueryBuilderConfig from sentry.snuba.dataset import Dataset from sentry.testutils.cases import TestCase +from sentry.testutils.helpers import override_options from sentry.utils.snuba import OPERATOR_TO_FUNCTION [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield + + # Helper functions to make reading the expected output from the boolean tests easier to read. # # a:b def _eq(xy): diff --git a/tests/sentry/test_culprit.py b/tests/sentry/test_culprit.py index b03986c81877b1..5696c2c6a62aeb 100644 --- a/tests/sentry/test_culprit.py +++ b/tests/sentry/test_culprit.py @@ -1,5 +1,14 @@ +import pytest + from sentry.event_manager import EventManager from sentry.event_manager import get_culprit as get_culprit_impl +from sentry.testutils.helpers import override_options + + [email protected](autouse=True) +def run_before_each(): + with override_options({"relay.enable-orjson": 0.0}): + yield def get_culprit(data):
5e2f314a70b45b4943fcd65a4a81a69bf6b042ca
2023-07-13 18:02:33
Anton Pirker
fix: Errors in Crons show up in wrong monitor. (#52788)
false
Errors in Crons show up in wrong monitor. (#52788)
fix
diff --git a/requirements-base.txt b/requirements-base.txt index 436e206e51fcb1..df999322ace013 100644 --- a/requirements-base.txt +++ b/requirements-base.txt @@ -62,7 +62,7 @@ sentry-arroyo>=2.14.1 sentry-kafka-schemas>=0.1.16 sentry-redis-tools>=0.1.7 sentry-relay>=0.8.28 -sentry-sdk>=1.28.0 +sentry-sdk>=1.28.1 snuba-sdk>=1.0.5 simplejson>=3.17.6 sqlparse>=0.4.4 diff --git a/requirements-dev-frozen.txt b/requirements-dev-frozen.txt index 800d64027d7073..66150596ed2df2 100644 --- a/requirements-dev-frozen.txt +++ b/requirements-dev-frozen.txt @@ -172,7 +172,7 @@ sentry-cli==2.16.0 sentry-kafka-schemas==0.1.16 sentry-redis-tools==0.1.7 sentry-relay==0.8.28 -sentry-sdk==1.28.0 +sentry-sdk==1.28.1 simplejson==3.17.6 six==1.16.0 sniffio==1.2.0 diff --git a/requirements-frozen.txt b/requirements-frozen.txt index a89449d0f28d74..c4af395218369d 100644 --- a/requirements-frozen.txt +++ b/requirements-frozen.txt @@ -119,7 +119,7 @@ sentry-arroyo==2.14.1 sentry-kafka-schemas==0.1.16 sentry-redis-tools==0.1.7 sentry-relay==0.8.28 -sentry-sdk==1.28.0 +sentry-sdk==1.28.1 simplejson==3.17.6 six==1.16.0 sniffio==1.2.0
5d4b5d62853fefd0cdc61828da26ca94d10d401e
2024-08-14 21:46:00
Luca Forstner
feat(webpack): Add Sentry Webpack Plugin (#76051)
false
Add Sentry Webpack Plugin (#76051)
feat
diff --git a/babel.config.ts b/babel.config.ts index 19214a68434db6..60efcd83e450a3 100644 --- a/babel.config.ts +++ b/babel.config.ts @@ -25,16 +25,12 @@ const config: TransformOptions = { plugins: ['@emotion/babel-plugin', '@babel/plugin-transform-runtime'], env: { production: { - plugins: [ - ['babel-plugin-add-react-displayname'], - '@sentry/babel-plugin-component-annotate', - ], + plugins: [['babel-plugin-add-react-displayname']], }, development: { plugins: [ '@emotion/babel-plugin', '@babel/plugin-transform-react-jsx-source', - '@sentry/babel-plugin-component-annotate', ...(process.env.SENTRY_UI_HOT_RELOAD ? ['react-refresh/babel'] : []), ], }, diff --git a/package.json b/package.json index efa272742ded15..6c566e097ecfe3 100644 --- a/package.json +++ b/package.json @@ -58,7 +58,6 @@ "@sentry-internal/rrweb": "2.25.0", "@sentry-internal/rrweb-player": "2.25.0", "@sentry-internal/rrweb-snapshot": "2.25.0", - "@sentry/babel-plugin-component-annotate": "^2.16.1", "@sentry/core": "^8.24.0", "@sentry/node": "^8.24.0", "@sentry/react": "^8.24.0", @@ -66,6 +65,7 @@ "@sentry/status-page-list": "^0.3.0", "@sentry/types": "^8.24.0", "@sentry/utils": "^8.24.0", + "@sentry/webpack-plugin": "^2.22.2", "@spotlightjs/spotlight": "^2.0.0-alpha.1", "@tanstack/react-query": "^4.29.7", "@tanstack/react-query-devtools": "^4.36.1", @@ -263,9 +263,7 @@ "last 3 iOS major versions", "Firefox ESR" ], - "test": [ - "current node" - ] + "test": ["current node"] }, "volta": { "extends": ".volta.json" diff --git a/static/app/bootstrap/initializeSdk.tsx b/static/app/bootstrap/initializeSdk.tsx index a1c09c2dc1d49f..d7669e84e955a3 100644 --- a/static/app/bootstrap/initializeSdk.tsx +++ b/static/app/bootstrap/initializeSdk.tsx @@ -64,6 +64,10 @@ function getSentryIntegrations(routes?: Function) { }, }), Sentry.browserProfilingIntegration(), + Sentry.thirdPartyErrorFilterIntegration({ + filterKeys: ['sentry-spa'], + behaviour: 'apply-tag-if-contains-third-party-frames', + }), ]; return integrations; diff --git a/webpack.config.ts b/webpack.config.ts index be9626e47af738..3dbcf93a3324ad 100644 --- a/webpack.config.ts +++ b/webpack.config.ts @@ -2,6 +2,7 @@ import {WebpackReactSourcemapsPlugin} from '@acemarke/react-prod-sourcemaps'; import {RsdoctorWebpackPlugin} from '@rsdoctor/webpack-plugin'; +import {sentryWebpackPlugin} from '@sentry/webpack-plugin'; import browserslist from 'browserslist'; import CompressionPlugin from 'compression-webpack-plugin'; import CopyPlugin from 'copy-webpack-plugin'; @@ -812,4 +813,23 @@ if (env.WEBPACK_CACHE_PATH) { }; } +appConfig.plugins?.push( + sentryWebpackPlugin({ + applicationKey: 'sentry-spa', + telemetry: false, + sourcemaps: { + disable: true, + }, + release: { + create: false, + }, + reactComponentAnnotation: { + enabled: true, + }, + bundleSizeOptimizations: { + excludeDebugStatements: IS_PRODUCTION, + }, + }) +); + export default appConfig; diff --git a/yarn.lock b/yarn.lock index d328e48eb61e87..103d64a915a0cb 100644 --- a/yarn.lock +++ b/yarn.lock @@ -139,7 +139,7 @@ "@babel/highlight" "^7.24.2" picocolors "^1.0.0" -"@babel/code-frame@^7.0.0", "@babel/code-frame@^7.10.4", "@babel/code-frame@^7.12.13", "@babel/code-frame@^7.16.7", "@babel/code-frame@^7.23.5", "@babel/code-frame@^7.24.2": +"@babel/code-frame@^7.0.0", "@babel/code-frame@^7.10.4", "@babel/code-frame@^7.12.13", "@babel/code-frame@^7.16.7", "@babel/code-frame@^7.23.5", "@babel/code-frame@^7.24.2", "@babel/code-frame@^7.24.7": version "7.24.7" resolved "https://registry.yarnpkg.com/@babel/code-frame/-/code-frame-7.24.7.tgz#882fd9e09e8ee324e496bd040401c6f046ef4465" integrity sha512-BcYH1CVJBO9tvyIZ2jVeXgSIMvGZ2FDRvDdOIVQyuklNKSsx+eppDEBq/g47Ayw+RqNFE+URvOShmf+f/qwAlA== @@ -152,6 +152,11 @@ resolved "https://registry.yarnpkg.com/@babel/compat-data/-/compat-data-7.24.4.tgz#6f102372e9094f25d908ca0d34fc74c74606059a" integrity sha512-vg8Gih2MLK+kOkHJp4gBEIkyaIi00jgWot2D9QOmmfLC8jINSOzmCLta6Bvz/JSBCqnegV0L80jhxkol5GWNfQ== +"@babel/compat-data@^7.25.2": + version "7.25.2" + resolved "https://registry.yarnpkg.com/@babel/compat-data/-/compat-data-7.25.2.tgz#e41928bd33475305c586f6acbbb7e3ade7a6f7f5" + integrity sha512-bYcppcpKBvX4znYaPEeFau03bp89ShqNMLs+rmdptMw+heSZh9+z84d2YG+K7cYLbWwzdjtDoW/uqZmPjulClQ== + "@babel/core@^7.0.0", "@babel/core@^7.11.6", "@babel/core@^7.12.3", "@babel/core@^7.23.9", "@babel/core@~7.24.5": version "7.24.5" resolved "https://registry.yarnpkg.com/@babel/core/-/core-7.24.5.tgz#15ab5b98e101972d171aeef92ac70d8d6718f06a" @@ -173,6 +178,27 @@ json5 "^2.2.3" semver "^6.3.1" +"@babel/core@^7.18.5": + version "7.25.2" + resolved "https://registry.yarnpkg.com/@babel/core/-/core-7.25.2.tgz#ed8eec275118d7613e77a352894cd12ded8eba77" + integrity sha512-BBt3opiCOxUr9euZ5/ro/Xv8/V7yJ5bjYMqG/C1YAo8MIKAnumZalCN+msbci3Pigy4lIQfPUpfMM27HMGaYEA== + dependencies: + "@ampproject/remapping" "^2.2.0" + "@babel/code-frame" "^7.24.7" + "@babel/generator" "^7.25.0" + "@babel/helper-compilation-targets" "^7.25.2" + "@babel/helper-module-transforms" "^7.25.2" + "@babel/helpers" "^7.25.0" + "@babel/parser" "^7.25.0" + "@babel/template" "^7.25.0" + "@babel/traverse" "^7.25.2" + "@babel/types" "^7.25.2" + convert-source-map "^2.0.0" + debug "^4.1.0" + gensync "^1.0.0-beta.2" + json5 "^2.2.3" + semver "^6.3.1" + "@babel/generator@^7.24.5", "@babel/generator@^7.7.2": version "7.24.5" resolved "https://registry.yarnpkg.com/@babel/generator/-/generator-7.24.5.tgz#e5afc068f932f05616b66713e28d0f04e99daeb3" @@ -183,6 +209,16 @@ "@jridgewell/trace-mapping" "^0.3.25" jsesc "^2.5.1" +"@babel/generator@^7.25.0": + version "7.25.0" + resolved "https://registry.yarnpkg.com/@babel/generator/-/generator-7.25.0.tgz#f858ddfa984350bc3d3b7f125073c9af6988f18e" + integrity sha512-3LEEcj3PVW8pW2R1SR1M89g/qrYk/m/mB/tLqn7dn4sbBUQyTqnlod+II2U4dqiGtUmkcnAmkMDralTFZttRiw== + dependencies: + "@babel/types" "^7.25.0" + "@jridgewell/gen-mapping" "^0.3.5" + "@jridgewell/trace-mapping" "^0.3.25" + jsesc "^2.5.1" + "@babel/helper-annotate-as-pure@^7.22.5": version "7.22.5" resolved "https://registry.yarnpkg.com/@babel/helper-annotate-as-pure/-/helper-annotate-as-pure-7.22.5.tgz#e7f06737b197d580a01edf75d97e2c8be99d3882" @@ -208,6 +244,17 @@ lru-cache "^5.1.1" semver "^6.3.1" +"@babel/helper-compilation-targets@^7.25.2": + version "7.25.2" + resolved "https://registry.yarnpkg.com/@babel/helper-compilation-targets/-/helper-compilation-targets-7.25.2.tgz#e1d9410a90974a3a5a66e84ff55ef62e3c02d06c" + integrity sha512-U2U5LsSaZ7TAt3cfaymQ8WHh0pxvdHoEk6HVpaexxixjyEquMh0L0YNJNM6CTGKMXV1iksi0iZkGw4AcFkPaaw== + dependencies: + "@babel/compat-data" "^7.25.2" + "@babel/helper-validator-option" "^7.24.8" + browserslist "^4.23.1" + lru-cache "^5.1.1" + semver "^6.3.1" + "@babel/helper-create-class-features-plugin@^7.24.1", "@babel/helper-create-class-features-plugin@^7.24.4", "@babel/helper-create-class-features-plugin@^7.24.5": version "7.24.5" resolved "https://registry.yarnpkg.com/@babel/helper-create-class-features-plugin/-/helper-create-class-features-plugin-7.24.5.tgz#7d19da92c7e0cd8d11c09af2ce1b8e7512a6e723" @@ -277,6 +324,14 @@ dependencies: "@babel/types" "^7.24.0" +"@babel/helper-module-imports@^7.24.7": + version "7.24.7" + resolved "https://registry.yarnpkg.com/@babel/helper-module-imports/-/helper-module-imports-7.24.7.tgz#f2f980392de5b84c3328fc71d38bd81bbb83042b" + integrity sha512-8AyH3C+74cgCVVXow/myrynrAGv+nTVg5vKu2nZph9x7RcRwzmh0VFallJuFTZ9mx6u4eSdXZfcOzSqTUm0HCA== + dependencies: + "@babel/traverse" "^7.24.7" + "@babel/types" "^7.24.7" + "@babel/helper-module-transforms@^7.23.3", "@babel/helper-module-transforms@^7.24.5": version "7.24.5" resolved "https://registry.yarnpkg.com/@babel/helper-module-transforms/-/helper-module-transforms-7.24.5.tgz#ea6c5e33f7b262a0ae762fd5986355c45f54a545" @@ -288,6 +343,16 @@ "@babel/helper-split-export-declaration" "^7.24.5" "@babel/helper-validator-identifier" "^7.24.5" +"@babel/helper-module-transforms@^7.25.2": + version "7.25.2" + resolved "https://registry.yarnpkg.com/@babel/helper-module-transforms/-/helper-module-transforms-7.25.2.tgz#ee713c29768100f2776edf04d4eb23b8d27a66e6" + integrity sha512-BjyRAbix6j/wv83ftcVJmBt72QtHI56C7JXZoG2xATiLpmoC7dpd8WnkikExHDVPpi/3qCmO6WY1EaXOluiecQ== + dependencies: + "@babel/helper-module-imports" "^7.24.7" + "@babel/helper-simple-access" "^7.24.7" + "@babel/helper-validator-identifier" "^7.24.7" + "@babel/traverse" "^7.25.2" + "@babel/helper-optimise-call-expression@^7.22.5": version "7.22.5" resolved "https://registry.yarnpkg.com/@babel/helper-optimise-call-expression/-/helper-optimise-call-expression-7.22.5.tgz#f21531a9ccbff644fdd156b4077c16ff0c3f609e" @@ -325,6 +390,14 @@ dependencies: "@babel/types" "^7.24.5" +"@babel/helper-simple-access@^7.24.7": + version "7.24.7" + resolved "https://registry.yarnpkg.com/@babel/helper-simple-access/-/helper-simple-access-7.24.7.tgz#bcade8da3aec8ed16b9c4953b74e506b51b5edb3" + integrity sha512-zBAIvbCMh5Ts+b86r/CjU+4XGYIs+R1j951gxI3KmmxBMhCg4oQMsv6ZXQ64XOm/cvzfU1FmoCyt6+owc5QMYg== + dependencies: + "@babel/traverse" "^7.24.7" + "@babel/types" "^7.24.7" + "@babel/helper-skip-transparent-expression-wrappers@^7.22.5": version "7.22.5" resolved "https://registry.yarnpkg.com/@babel/helper-skip-transparent-expression-wrappers/-/helper-skip-transparent-expression-wrappers-7.22.5.tgz#007f15240b5751c537c40e77abb4e89eeaaa8847" @@ -344,6 +417,11 @@ resolved "https://registry.yarnpkg.com/@babel/helper-string-parser/-/helper-string-parser-7.24.1.tgz#f99c36d3593db9540705d0739a1f10b5e20c696e" integrity sha512-2ofRCjnnA9y+wk8b9IAREroeUP02KHp431N2mhKniy2yKIDKpbrHv9eXwm8cBeWQYcJmzv5qKCu65P47eCF7CQ== +"@babel/helper-string-parser@^7.24.8": + version "7.24.8" + resolved "https://registry.yarnpkg.com/@babel/helper-string-parser/-/helper-string-parser-7.24.8.tgz#5b3329c9a58803d5df425e5785865881a81ca48d" + integrity sha512-pO9KhhRcuUyGnJWwyEgnRJTSIZHiT+vMD0kPeD+so0l7mxkMT19g3pjY9GTnHySck/hDzq+dtW/4VgnMkippsQ== + "@babel/helper-validator-identifier@^7.22.20", "@babel/helper-validator-identifier@^7.24.5", "@babel/helper-validator-identifier@^7.24.7": version "7.24.7" resolved "https://registry.yarnpkg.com/@babel/helper-validator-identifier/-/helper-validator-identifier-7.24.7.tgz#75b889cfaf9e35c2aaf42cf0d72c8e91719251db" @@ -354,6 +432,11 @@ resolved "https://registry.yarnpkg.com/@babel/helper-validator-option/-/helper-validator-option-7.23.5.tgz#907a3fbd4523426285365d1206c423c4c5520307" integrity sha512-85ttAOMLsr53VgXkTbkx8oA6YTfT4q7/HzXSLEYmjcSTJPMPQtvq1BD79Byep5xMUYbGRzEpDsjUf3dyp54IKw== +"@babel/helper-validator-option@^7.24.8": + version "7.24.8" + resolved "https://registry.yarnpkg.com/@babel/helper-validator-option/-/helper-validator-option-7.24.8.tgz#3725cdeea8b480e86d34df15304806a06975e33d" + integrity sha512-xb8t9tD1MHLungh/AIoWYN+gVHaB9kwlu8gffXGSt3FFEIT7RjS+xWbc2vUD1UTZdIpKj/ab3rdqJ7ufngyi2Q== + "@babel/helper-wrap-function@^7.22.20": version "7.24.5" resolved "https://registry.yarnpkg.com/@babel/helper-wrap-function/-/helper-wrap-function-7.24.5.tgz#335f934c0962e2c1ed1fb9d79e06a56115067c09" @@ -372,6 +455,14 @@ "@babel/traverse" "^7.24.5" "@babel/types" "^7.24.5" +"@babel/helpers@^7.25.0": + version "7.25.0" + resolved "https://registry.yarnpkg.com/@babel/helpers/-/helpers-7.25.0.tgz#e69beb7841cb93a6505531ede34f34e6a073650a" + integrity sha512-MjgLZ42aCm0oGjJj8CtSM3DB8NOOf8h2l7DCTePJs29u+v7yO/RBX9nShlKMgFnRks/Q4tBAe7Hxnov9VkGwLw== + dependencies: + "@babel/template" "^7.25.0" + "@babel/types" "^7.25.0" + "@babel/highlight@^7.24.2", "@babel/highlight@^7.24.7": version "7.24.7" resolved "https://registry.yarnpkg.com/@babel/highlight/-/highlight-7.24.7.tgz#a05ab1df134b286558aae0ed41e6c5f731bf409d" @@ -387,6 +478,13 @@ resolved "https://registry.yarnpkg.com/@babel/parser/-/parser-7.24.5.tgz#4a4d5ab4315579e5398a82dcf636ca80c3392790" integrity sha512-EOv5IK8arwh3LI47dz1b0tKUb/1uhHAnHJOrjgtQMIpu1uXd9mlFrJg9IUgGUgZ41Ch0K8REPTYpO7B76b4vJg== +"@babel/parser@^7.25.0", "@babel/parser@^7.25.3": + version "7.25.3" + resolved "https://registry.yarnpkg.com/@babel/parser/-/parser-7.25.3.tgz#91fb126768d944966263f0657ab222a642b82065" + integrity sha512-iLTJKDbJ4hMvFPgQwwsVoxtHyWpKKPBrxkANrSYewDPaPpT5py5yeVkgPIJ7XYXhndxJpaA3PyALSXQ7u8e/Dw== + dependencies: + "@babel/types" "^7.25.2" + "@babel/plugin-bugfix-firefox-class-in-computed-class-key@^7.24.5": version "7.24.5" resolved "https://registry.yarnpkg.com/@babel/plugin-bugfix-firefox-class-in-computed-class-key/-/plugin-bugfix-firefox-class-in-computed-class-key-7.24.5.tgz#4c3685eb9cd790bcad2843900fe0250c91ccf895" @@ -1161,6 +1259,15 @@ "@babel/parser" "^7.24.0" "@babel/types" "^7.24.0" +"@babel/template@^7.25.0": + version "7.25.0" + resolved "https://registry.yarnpkg.com/@babel/template/-/template-7.25.0.tgz#e733dc3134b4fede528c15bc95e89cb98c52592a" + integrity sha512-aOOgh1/5XzKvg1jvVz7AVrx2piJ2XBi227DHmbY6y+bM9H2FlN+IfecYu4Xl0cNiiVejlsCri89LUsbj8vJD9Q== + dependencies: + "@babel/code-frame" "^7.24.7" + "@babel/parser" "^7.25.0" + "@babel/types" "^7.25.0" + "@babel/traverse@^7.24.5": version "7.24.5" resolved "https://registry.yarnpkg.com/@babel/traverse/-/traverse-7.24.5.tgz#972aa0bc45f16983bf64aa1f877b2dd0eea7e6f8" @@ -1177,6 +1284,19 @@ debug "^4.3.1" globals "^11.1.0" +"@babel/traverse@^7.24.7", "@babel/traverse@^7.25.2": + version "7.25.3" + resolved "https://registry.yarnpkg.com/@babel/traverse/-/traverse-7.25.3.tgz#f1b901951c83eda2f3e29450ce92743783373490" + integrity sha512-HefgyP1x754oGCsKmV5reSmtV7IXj/kpaE1XYY+D9G5PvKKoFfSbiS4M77MdjuwlZKDIKFCffq9rPU+H/s3ZdQ== + dependencies: + "@babel/code-frame" "^7.24.7" + "@babel/generator" "^7.25.0" + "@babel/parser" "^7.25.3" + "@babel/template" "^7.25.0" + "@babel/types" "^7.25.2" + debug "^4.3.1" + globals "^11.1.0" + "@babel/types@^7.0.0", "@babel/types@^7.22.15", "@babel/types@^7.22.5", "@babel/types@^7.23.0", "@babel/types@^7.23.4", "@babel/types@^7.24.0", "@babel/types@^7.24.5", "@babel/types@^7.3.0", "@babel/types@^7.3.3", "@babel/types@^7.4.4": version "7.24.5" resolved "https://registry.yarnpkg.com/@babel/types/-/types-7.24.5.tgz#7661930afc638a5383eb0c4aee59b74f38db84d7" @@ -1186,6 +1306,15 @@ "@babel/helper-validator-identifier" "^7.24.5" to-fast-properties "^2.0.0" +"@babel/types@^7.24.7", "@babel/types@^7.25.0", "@babel/types@^7.25.2": + version "7.25.2" + resolved "https://registry.yarnpkg.com/@babel/types/-/types-7.25.2.tgz#55fb231f7dc958cd69ea141a4c2997e819646125" + integrity sha512-YTnYtra7W9e6/oAZEHj0bJehPRUlLH9/fbpT5LfB0NhQXyALCRkRs3zH9v07IYhkgpqX6Z78FnuccZr/l4Fs4Q== + dependencies: + "@babel/helper-string-parser" "^7.24.8" + "@babel/helper-validator-identifier" "^7.24.7" + to-fast-properties "^2.0.0" + "@bcoe/v8-coverage@^0.2.3": version "0.2.3" resolved "https://registry.yarnpkg.com/@bcoe/v8-coverage/-/v8-coverage-0.2.3.tgz#75a2e8b51cb758a7553d6804a5932d7aace75c39" @@ -3124,10 +3253,10 @@ fflate "^0.4.4" mitt "^3.0.0" -"@sentry/babel-plugin-component-annotate@^2.16.1": - version "2.16.1" - resolved "https://registry.yarnpkg.com/@sentry/babel-plugin-component-annotate/-/babel-plugin-component-annotate-2.16.1.tgz#da3bf4ec1c1dc68a97d6a7e27bd710001d6b07fb" - integrity sha512-pJka66URsqQbk6hTs9H1XFpUeI0xxuqLYf9Dy5pRGNHSJMtfv91U+CaYSWt03aRRMGDXMduh62zAAY7Wf0HO+A== +"@sentry/[email protected]": + version "2.22.2" + resolved "https://registry.yarnpkg.com/@sentry/babel-plugin-component-annotate/-/babel-plugin-component-annotate-2.22.2.tgz#f4a1ddea4bcac06584a6cec9a43ec088cbb6caaf" + integrity sha512-6kFAHGcs0npIC4HTt4ULs8uOfEucvMI7VW4hoyk17jhRaW8CbxzxfWCfIeRbDkE8pYwnARaq83tu025Hrk2zgA== "@sentry/[email protected]": version "8.24.0" @@ -3142,6 +3271,74 @@ "@sentry/types" "8.24.0" "@sentry/utils" "8.24.0" +"@sentry/[email protected]": + version "2.22.2" + resolved "https://registry.yarnpkg.com/@sentry/bundler-plugin-core/-/bundler-plugin-core-2.22.2.tgz#bd418541245c5167a439d4e28a84096deb20c512" + integrity sha512-TwEEW4FeEJ5Mamp4fGnktfVjzN77KAW0xFQsEPuxZtOAPG17zX/PGvdyRX/TE1jkZWhTzqUDIdgzqlNLjyEnUw== + dependencies: + "@babel/core" "^7.18.5" + "@sentry/babel-plugin-component-annotate" "2.22.2" + "@sentry/cli" "^2.33.1" + dotenv "^16.3.1" + find-up "^5.0.0" + glob "^9.3.2" + magic-string "0.30.8" + unplugin "1.0.1" + +"@sentry/[email protected]": + version "2.33.1" + resolved "https://registry.yarnpkg.com/@sentry/cli-darwin/-/cli-darwin-2.33.1.tgz#e4eb1dd01ee3ce2788025426b860ccc63759589c" + integrity sha512-+4/VIx/E1L2hChj5nGf5MHyEPHUNHJ/HoG5RY+B+vyEutGily1c1+DM2bum7RbD0xs6wKLIyup5F02guzSzG8A== + +"@sentry/[email protected]": + version "2.33.1" + resolved "https://registry.yarnpkg.com/@sentry/cli-linux-arm64/-/cli-linux-arm64-2.33.1.tgz#9ea1718c21ef32ca83b0852ca29fb461fd26d25a" + integrity sha512-DbGV56PRKOLsAZJX27Jt2uZ11QfQEMmWB4cIvxkKcFVE+LJP4MVA+MGGRUL6p+Bs1R9ZUuGbpKGtj0JiG6CoXw== + +"@sentry/[email protected]": + version "2.33.1" + resolved "https://registry.yarnpkg.com/@sentry/cli-linux-arm/-/cli-linux-arm-2.33.1.tgz#e8a1dca4d008dd6a72ab5935304c104e98e2901c" + integrity sha512-zbxEvQju+tgNvzTOt635le4kS/Fbm2XC2RtYbCTs034Vb8xjrAxLnK0z1bQnStUV8BkeBHtsNVrG+NSQDym2wg== + +"@sentry/[email protected]": + version "2.33.1" + resolved "https://registry.yarnpkg.com/@sentry/cli-linux-i686/-/cli-linux-i686-2.33.1.tgz#f1fe8dd4d6dde0812a94fba31de8054ddfb7284a" + integrity sha512-g2LS4oPXkPWOfKWukKzYp4FnXVRRSwBxhuQ9eSw2peeb58ZIObr4YKGOA/8HJRGkooBJIKGaAR2mH2Pk1TKaiA== + +"@sentry/[email protected]": + version "2.33.1" + resolved "https://registry.yarnpkg.com/@sentry/cli-linux-x64/-/cli-linux-x64-2.33.1.tgz#6e086675356a9eb79731bf9e447d078bae1b5adf" + integrity sha512-IV3dcYV/ZcvO+VGu9U6kuxSdbsV2kzxaBwWUQxtzxJ+cOa7J8Hn1t0koKGtU53JVZNBa06qJWIcqgl4/pCuKIg== + +"@sentry/[email protected]": + version "2.33.1" + resolved "https://registry.yarnpkg.com/@sentry/cli-win32-i686/-/cli-win32-i686-2.33.1.tgz#0e6b36c4a2f5f6e85a59247a123d276b3ef10f1a" + integrity sha512-F7cJySvkpzIu7fnLKNHYwBzZYYwlhoDbAUnaFX0UZCN+5DNp/5LwTp37a5TWOsmCaHMZT4i9IO4SIsnNw16/zQ== + +"@sentry/[email protected]": + version "2.33.1" + resolved "https://registry.yarnpkg.com/@sentry/cli-win32-x64/-/cli-win32-x64-2.33.1.tgz#2d00b38a2dd9f3355df91825582ada3ea0034e86" + integrity sha512-8VyRoJqtb2uQ8/bFRKNuACYZt7r+Xx0k2wXRGTyH05lCjAiVIXn7DiS2BxHFty7M1QEWUCMNsb/UC/x/Cu2wuA== + +"@sentry/cli@^2.33.1": + version "2.33.1" + resolved "https://registry.yarnpkg.com/@sentry/cli/-/cli-2.33.1.tgz#cfbdffdd896b05b92a659baf435b5607037af928" + integrity sha512-dUlZ4EFh98VFRPJ+f6OW3JEYQ7VvqGNMa0AMcmvk07ePNeK/GicAWmSQE4ZfJTTl80ul6HZw1kY01fGQOQlVRA== + dependencies: + https-proxy-agent "^5.0.0" + node-fetch "^2.6.7" + progress "^2.0.3" + proxy-from-env "^1.1.0" + which "^2.0.2" + optionalDependencies: + "@sentry/cli-darwin" "2.33.1" + "@sentry/cli-linux-arm" "2.33.1" + "@sentry/cli-linux-arm64" "2.33.1" + "@sentry/cli-linux-i686" "2.33.1" + "@sentry/cli-linux-x64" "2.33.1" + "@sentry/cli-win32-i686" "2.33.1" + "@sentry/cli-win32-x64" "2.33.1" + "@sentry/[email protected]", "@sentry/core@^8.24.0": version "8.24.0" resolved "https://registry.yarnpkg.com/@sentry/core/-/core-8.24.0.tgz#63fd159c69e897e3a3d1c260e49824f7d444352a" @@ -3245,6 +3442,15 @@ dependencies: "@sentry/types" "8.24.0" +"@sentry/webpack-plugin@^2.22.2": + version "2.22.2" + resolved "https://registry.yarnpkg.com/@sentry/webpack-plugin/-/webpack-plugin-2.22.2.tgz#576de848985d29c524370f8da623bcaf19bf16df" + integrity sha512-zVPs3BLClHM8jIjr2FChux16GMLel8OjXBjd4V8/r1Kf2fGiQDXo72GxsrW8AdVlIHgQApLzubuQ2kpcFVK4Sw== + dependencies: + "@sentry/bundler-plugin-core" "2.22.2" + unplugin "1.0.1" + uuid "^9.0.0" + "@sinclair/typebox@^0.27.8": version "0.27.8" resolved "https://registry.yarnpkg.com/@sinclair/typebox/-/typebox-0.27.8.tgz#6667fac16c436b5434a387a34dedb013198f6e6e" @@ -4359,6 +4565,11 @@ acorn@^8.0.4, acorn@^8.1.0, acorn@^8.10.0, acorn@^8.11.0, acorn@^8.11.3, acorn@^ resolved "https://registry.yarnpkg.com/acorn/-/acorn-8.12.0.tgz#1627bfa2e058148036133b8d9b51a700663c294c" integrity sha512-RTvkC4w+KNXrM39/lWCUaG0IbRkWdCv7W/IOW9oU6SawyxulvkQy5HQPVTKxEjczcUvapcrw3cFx/60VN/NRNw== +acorn@^8.8.1: + version "8.12.1" + resolved "https://registry.yarnpkg.com/acorn/-/acorn-8.12.1.tgz#71616bdccbe25e27a54439e0046e89ca76df2248" + integrity sha512-tcpGyI9zbizT9JbV6oYE477V6mTlXvvi0T0G3SNIYE2apm/G5huBa1+K89VGeovbg+jycCrfhl3ADxErOuO6Jg== + agent-base@6: version "6.0.2" resolved "https://registry.yarnpkg.com/agent-base/-/agent-base-6.0.2.tgz#49fff58577cfee3f37176feab4c22e00f86d7f77" @@ -4909,6 +5120,16 @@ browserslist@^4.0.0, browserslist@^4.18.1, browserslist@^4.21.10, browserslist@^ node-releases "^2.0.14" update-browserslist-db "^1.0.16" +browserslist@^4.23.1: + version "4.23.3" + resolved "https://registry.yarnpkg.com/browserslist/-/browserslist-4.23.3.tgz#debb029d3c93ebc97ffbc8d9cbb03403e227c800" + integrity sha512-btwCFJVjI4YWDNfau8RhZ+B1Q/VLoUITrm3RlP6y1tYGWIOa+InuYiRGXUBXo8nA1qKmHMyLB/iVQg5TT4eFoA== + dependencies: + caniuse-lite "^1.0.30001646" + electron-to-chromium "^1.5.4" + node-releases "^2.0.18" + update-browserslist-db "^1.1.0" + [email protected]: version "2.1.1" resolved "https://registry.yarnpkg.com/bser/-/bser-2.1.1.tgz#e6787da20ece9d07998533cfd9de6f5c38f4bc05" @@ -4995,6 +5216,11 @@ caniuse-lite@^1.0.0, caniuse-lite@^1.0.30001283, caniuse-lite@^1.0.30001629: resolved "https://registry.yarnpkg.com/caniuse-lite/-/caniuse-lite-1.0.30001638.tgz#598e1f0c2ac36f37ebc3f5b8887a32ca558e5d56" integrity sha512-5SuJUJ7cZnhPpeLHaH0c/HPAnAHZvS6ElWyHK9GSIbVOQABLzowiI2pjmpvZ1WEbkyz46iFd4UXlOHR5SqgfMQ== +caniuse-lite@^1.0.30001646: + version "1.0.30001651" + resolved "https://registry.yarnpkg.com/caniuse-lite/-/caniuse-lite-1.0.30001651.tgz#52de59529e8b02b1aedcaaf5c05d9e23c0c28138" + integrity sha512-9Cf+Xv1jJNe1xPZLGuUXLNkE1BoDkqRqYyFJ9TDYSqhduqA4hu4oR9HluGoWYQC/aj8WHjsGVV+bwkh0+tegRg== + cbor-web@^8.1.0: version "8.1.0" resolved "https://registry.yarnpkg.com/cbor-web/-/cbor-web-8.1.0.tgz#c1148e91ca6bfc0f5c07c1df164854596e2e33d6" @@ -5943,6 +6169,11 @@ dot-case@^3.0.4: no-case "^3.0.4" tslib "^2.0.3" +dotenv@^16.3.1: + version "16.4.5" + resolved "https://registry.yarnpkg.com/dotenv/-/dotenv-16.4.5.tgz#cdd3b3b604cb327e286b4762e13502f717cb099f" + integrity sha512-ZmdL2rui+eB2YwhsWzjInR8LldtZHGDoQ1ugH85ppHKwpUHL7j7rN0Ti9NCnGiQbhaZ11FpR+7ao1dNsmduNUg== + [email protected]: version "1.4.0" resolved "https://registry.yarnpkg.com/downsample/-/downsample-1.4.0.tgz#f7500318efd5d979641d45fe49cd2dc8d7cb9397" @@ -5994,6 +6225,11 @@ electron-to-chromium@^1.4.796: resolved "https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.4.812.tgz#21b78709c5a13af5d5c688d135a22dcea7617acf" integrity sha512-7L8fC2Ey/b6SePDFKR2zHAy4mbdp1/38Yk5TsARO66W3hC5KEaeKMMHoxwtuH+jcu2AYLSn9QX04i95t6Fl1Hg== +electron-to-chromium@^1.5.4: + version "1.5.6" + resolved "https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.5.6.tgz#c81d9938b5a877314ad370feb73b4e5409b36abd" + integrity sha512-jwXWsM5RPf6j9dPYzaorcBSUg6AiqocPEyMpkchkvntaH9HGfOOMZwxMJjDY/XEs3T5dM7uyH1VhRMkqUU9qVw== + emittery@^0.13.1: version "0.13.1" resolved "https://registry.yarnpkg.com/emittery/-/emittery-0.13.1.tgz#c04b8c3457490e0847ae51fced3af52d338e3dad" @@ -7090,6 +7326,16 @@ glob@^7.1.3, glob@^7.1.4: once "^1.3.0" path-is-absolute "^1.0.0" +glob@^9.3.2: + version "9.3.5" + resolved "https://registry.yarnpkg.com/glob/-/glob-9.3.5.tgz#ca2ed8ca452781a3009685607fdf025a899dfe21" + integrity sha512-e1LleDykUz2Iu+MTYdkSsuWX8lvAjAcs0Xef0lNIu0S2wOAzuTxCJtcd9S3cijlwYF18EsU3rzb8jPVobxDh9Q== + dependencies: + fs.realpath "^1.0.0" + minimatch "^8.0.2" + minipass "^4.2.4" + path-scurry "^1.6.1" + global-modules@^2.0.0: version "2.0.0" resolved "https://registry.yarnpkg.com/global-modules/-/global-modules-2.0.0.tgz#997605ad2345f27f51539bea26574421215c7780" @@ -7388,7 +7634,7 @@ http-proxy@^1.18.1: follow-redirects "^1.0.0" requires-port "^1.0.0" -https-proxy-agent@^5.0.1: +https-proxy-agent@^5.0.0, https-proxy-agent@^5.0.1: version "5.0.1" resolved "https://registry.yarnpkg.com/https-proxy-agent/-/https-proxy-agent-5.0.1.tgz#c59ef224a04fe8b754f3db0063a25ea30d0005d6" integrity sha512-dFcAjpTQFgoLMzC2VwU+C/CbS7uRL0lWmxDITmqm7C+7F0Odmj6s9l6alZc6AELXhrnggM2CeWSXHGOdX2YtwA== @@ -8764,6 +9010,11 @@ lower-case@^2.0.2: dependencies: tslib "^2.0.3" +lru-cache@^10.2.0: + version "10.4.3" + resolved "https://registry.yarnpkg.com/lru-cache/-/lru-cache-10.4.3.tgz#410fc8a17b70e598013df257c2446b7f3383f119" + integrity sha512-JNAzZcXrCt42VGLuYz0zfAzDfAvJWW6AfYlDBQyDV5DClI2m5sAmK+OIO7s59XfsRsWHp02jAJrRadPRGTt6SQ== + lru-cache@^5.1.1: version "5.1.1" resolved "https://registry.yarnpkg.com/lru-cache/-/lru-cache-5.1.1.tgz#1da27e6710271947695daf6848e847f01d84b920" @@ -8781,6 +9032,13 @@ lz-string@^1.5.0: resolved "https://registry.yarnpkg.com/lz-string/-/lz-string-1.5.0.tgz#c1ab50f77887b712621201ba9fd4e3a6ed099941" integrity sha512-h5bgJWpxJNswbU7qCrV0tIKQCaS3blPDrqKWx+QxzuzL1zGUzij9XCWLrSLsJPu5t+eWA/ycetzYAO5IOMcWAQ== [email protected]: + version "0.30.8" + resolved "https://registry.yarnpkg.com/magic-string/-/magic-string-0.30.8.tgz#14e8624246d2bedba70d5462aa99ac9681844613" + integrity sha512-ISQTe55T2ao7XtlAStud6qwYPZjE4GK1S/BeVPus4jrq6JuOnQ00YKQC581RWhR122W7msZV263KzVeLoqidyQ== + dependencies: + "@jridgewell/sourcemap-codec" "^1.4.15" + magic-string@^0.30.3: version "0.30.4" resolved "https://registry.yarnpkg.com/magic-string/-/magic-string-0.30.4.tgz#c2c683265fc18dda49b56fc7318d33ca0332c98c" @@ -8946,6 +9204,13 @@ minimatch@^3.0.4, minimatch@^3.0.5, minimatch@^3.1.1, minimatch@^3.1.2: dependencies: brace-expansion "^1.1.7" +minimatch@^8.0.2: + version "8.0.4" + resolved "https://registry.yarnpkg.com/minimatch/-/minimatch-8.0.4.tgz#847c1b25c014d4e9a7f68aaf63dedd668a626229" + integrity sha512-W0Wvr9HyFXZRGIDgCicunpQ299OKXs9RgZfaukz4qAW/pJhcpUfupc9c+OObPOFueNy8VSrZgEmDtk6Kh4WzDA== + dependencies: + brace-expansion "^2.0.1" + minimatch@^9.0.1, minimatch@^9.0.4: version "9.0.5" resolved "https://registry.yarnpkg.com/minimatch/-/minimatch-9.0.5.tgz#d74f9dd6b57d83d8e98cfb82133b03978bc929e5" @@ -8958,6 +9223,11 @@ minimist@^1.2.0, minimist@^1.2.6: resolved "https://registry.yarnpkg.com/minimist/-/minimist-1.2.8.tgz#c1a464e7693302e082a075cee0c057741ac4772c" integrity sha512-2yyAR8qBkN3YuheJanUpWC5U3bb5osDywNB8RzDVlDwDHbocAJveqqj1u8+SVD7jkWT4yvsHCpWqqWqAxb0zCA== +minipass@^4.2.4: + version "4.2.8" + resolved "https://registry.yarnpkg.com/minipass/-/minipass-4.2.8.tgz#f0010f64393ecfc1d1ccb5f582bcaf45f48e1a3a" + integrity sha512-fNzuVyifolSLFL4NzpF+wEF4qrgqaaKX0haXPQEdQ7NKAN+WecoKMHV09YcuL/DHxrUsYQOK3MiuDf7Ip2OXfQ== + "minipass@^5.0.0 || ^6.0.2 || ^7.0.0": version "7.0.4" resolved "https://registry.yarnpkg.com/minipass/-/minipass-7.0.4.tgz#dbce03740f50a4786ba994c1fb908844d27b038c" @@ -9101,6 +9371,13 @@ node-abort-controller@^3.0.1: resolved "https://registry.yarnpkg.com/node-abort-controller/-/node-abort-controller-3.0.1.tgz#f91fa50b1dee3f909afabb7e261b1e1d6b0cb74e" integrity sha512-/ujIVxthRs+7q6hsdjHMaj8hRG9NuWmwrz+JdRwZ14jdFoKSkm+vDsCbF9PLpnSqjaWQJuTmVtcWHNLr+vrOFw== +node-fetch@^2.6.7: + version "2.7.0" + resolved "https://registry.yarnpkg.com/node-fetch/-/node-fetch-2.7.0.tgz#d0f0fa6e3e2dc1d27efcd8ad99d550bda94d187d" + integrity sha512-c4FRfUm/dbcWZ7U+1Wq0AwCyFL+3nt2bEw05wfxSz+DWpWsitgmSgYmy2dQdWyKC1694ELPqMs/YzUSNozLt8A== + dependencies: + whatwg-url "^5.0.0" + node-forge@^1: version "1.3.1" resolved "https://registry.yarnpkg.com/node-forge/-/node-forge-1.3.1.tgz#be8da2af243b2417d5f646a770663a92b7e9ded3" @@ -9116,6 +9393,11 @@ node-releases@^2.0.14: resolved "https://registry.yarnpkg.com/node-releases/-/node-releases-2.0.14.tgz#2ffb053bceb8b2be8495ece1ab6ce600c4461b0b" integrity sha512-y10wOWt8yZpqXmOgRo77WaHEmhYQYGNA6y421PKsKYWEK8aW+cqAphborZDhqfyKrbZEN92CN1X2KbafY2s7Yw== +node-releases@^2.0.18: + version "2.0.18" + resolved "https://registry.yarnpkg.com/node-releases/-/node-releases-2.0.18.tgz#f010e8d35e2fe8d6b2944f03f70213ecedc4ca3f" + integrity sha512-d9VeXT4SJ7ZeOqGX6R5EM022wpL+eWPooLI+5UpWn2jCT1aosUQEhQP214x33Wkwx3JQMvIm+tIoVOdodFS40g== + nopt@^7.2.0: version "7.2.0" resolved "https://registry.yarnpkg.com/nopt/-/nopt-7.2.0.tgz#067378c68116f602f552876194fd11f1292503d7" @@ -9461,6 +9743,14 @@ path-scurry@^1.10.1: lru-cache "^9.1.1 || ^10.0.0" minipass "^5.0.0 || ^6.0.2 || ^7.0.0" +path-scurry@^1.6.1: + version "1.11.1" + resolved "https://registry.yarnpkg.com/path-scurry/-/path-scurry-1.11.1.tgz#7960a668888594a0720b12a911d1a742ab9f11d2" + integrity sha512-Xa4Nw17FS9ApQFJ9umLiJS4orGjm7ZzwUrwamcGQuHSzDyth9boKDaycYdDcZDuqYATXw4HFXgaqWTctW/v1HA== + dependencies: + lru-cache "^10.2.0" + minipass "^5.0.0 || ^6.0.2 || ^7.0.0" + [email protected]: version "0.1.7" resolved "https://registry.yarnpkg.com/path-to-regexp/-/path-to-regexp-0.1.7.tgz#df604178005f522f15eb4490e7247a1bfaa67f8c" @@ -9959,6 +10249,11 @@ process@^0.11.10: resolved "https://registry.yarnpkg.com/process/-/process-0.11.10.tgz#7332300e840161bda3e69a1d1d91a7d4bc16f182" integrity sha512-cdGef/drWFoydD1JsMzuFf8100nZl+GT+yacc2bEced5f9Rjk4z+WtFUTBu9PhOi9j/jfmBPu0mMEY4wIdAF8A== +progress@^2.0.3: + version "2.0.3" + resolved "https://registry.yarnpkg.com/progress/-/progress-2.0.3.tgz#7e8cf8d8f5b8f239c1bc68beb4eb78567d572ef8" + integrity sha512-7PiHtLll5LdnKIMw100I+8xJXR5gW2QwWYkT6iJva0bXitZKa/XMrSbdmg3r2Xnaidz9Qumd0VPaMrZlF9V9sA== + prompts@^2.0.1: version "2.4.0" resolved "https://registry.yarnpkg.com/prompts/-/prompts-2.4.0.tgz#4aa5de0723a231d1ee9121c40fdf663df73f61d7" @@ -11361,6 +11656,11 @@ tr46@^3.0.0: dependencies: punycode "^2.1.1" +tr46@~0.0.3: + version "0.0.3" + resolved "https://registry.yarnpkg.com/tr46/-/tr46-0.0.3.tgz#8184fd347dac9cdc185992f3a6622e14b9d9ab6a" + integrity sha512-N3WMsuqV66lT30CrXNbEjx4GEwlow3v6rr4mCcv6prnfwhS01rkgyFdjPNBYd9br7LpXV1+Emh01fHnq2Gdgrw== + ts-api-utils@^1.3.0: version "1.3.0" resolved "https://registry.yarnpkg.com/ts-api-utils/-/ts-api-utils-1.3.0.tgz#4b490e27129f1e8e686b45cc4ab63714dc60eea1" @@ -11580,6 +11880,16 @@ [email protected], unpipe@~1.0.0: resolved "https://registry.yarnpkg.com/unpipe/-/unpipe-1.0.0.tgz#b2bf4ee8514aae6165b4817829d21b2ef49904ec" integrity sha1-sr9O6FFKrmFltIF4KdIbLvSZBOw= [email protected]: + version "1.0.1" + resolved "https://registry.yarnpkg.com/unplugin/-/unplugin-1.0.1.tgz#83b528b981cdcea1cad422a12cd02e695195ef3f" + integrity sha512-aqrHaVBWW1JVKBHmGo33T5TxeL0qWzfvjWokObHA9bYmN7eNDkwOxmLjhioHl9878qDFMAaT51XNroRyuz7WxA== + dependencies: + acorn "^8.8.1" + chokidar "^3.5.3" + webpack-sources "^3.2.3" + webpack-virtual-modules "^0.5.0" + unplugin@^1.10.1, unplugin@^1.4.0: version "1.10.1" resolved "https://registry.yarnpkg.com/unplugin/-/unplugin-1.10.1.tgz#8ceda065dc71bc67d923dea0920f05c67f2cd68c" @@ -11598,6 +11908,14 @@ update-browserslist-db@^1.0.16: escalade "^3.1.2" picocolors "^1.0.1" +update-browserslist-db@^1.1.0: + version "1.1.0" + resolved "https://registry.yarnpkg.com/update-browserslist-db/-/update-browserslist-db-1.1.0.tgz#7ca61c0d8650766090728046e416a8cde682859e" + integrity sha512-EdRAaAyk2cUE1wOf2DkEhzxqOQvFOoRJFNS6NeyJ01Gp2beMRpBAINjM2iDXE3KCuKhwnvHIQCJm6ThL2Z+HzQ== + dependencies: + escalade "^3.1.2" + picocolors "^1.0.1" + uri-js@^4.2.2: version "4.4.1" resolved "https://registry.yarnpkg.com/uri-js/-/uri-js-4.4.1.tgz#9b1a52595225859e55f669d928f88c6c57f2a77e" @@ -11647,6 +11965,11 @@ uuid@^8.3.2: resolved "https://registry.yarnpkg.com/uuid/-/uuid-8.3.2.tgz#80d5b5ced271bb9af6c445f21a1a04c606cefbe2" integrity sha512-+NYs2QeMWy+GWFOEm9xnn6HCDp0l7QBD7ml8zLUmJ+93Q5NF0NocErnwkTkXVFNiX3/fpC6afS8Dhb/gz7R7eg== +uuid@^9.0.0: + version "9.0.1" + resolved "https://registry.yarnpkg.com/uuid/-/uuid-9.0.1.tgz#e188d4c8853cc722220392c424cd637f32293f30" + integrity sha512-b+1eJOlsR9K8HJpow9Ok3fiWOWSIcIzXodvv0rQjVoOVNpWMpxf1wZNpt4y9h10odCNrqnYp1OBzRktckBe3sA== + v8-compile-cache-lib@^3.0.1: version "3.0.1" resolved "https://registry.yarnpkg.com/v8-compile-cache-lib/-/v8-compile-cache-lib-3.0.1.tgz#6336e8d71965cb3d35a1bbb7868445a7c05264bf" @@ -11739,6 +12062,11 @@ wbuf@^1.1.0, wbuf@^1.7.3: dependencies: minimalistic-assert "^1.0.0" +webidl-conversions@^3.0.0: + version "3.0.1" + resolved "https://registry.yarnpkg.com/webidl-conversions/-/webidl-conversions-3.0.1.tgz#24534275e2a7bc6be7bc86611cc16ae0a5654871" + integrity sha512-2JAn3z8AR6rjK8Sm8orRC0h/bcl/DqL7tRPdGZ4I1CjdF+EaMLmYxBHyXuKL849eucPFhvBoxMsflfOb8kxaeQ== + webidl-conversions@^7.0.0: version "7.0.0" resolved "https://registry.yarnpkg.com/webidl-conversions/-/webidl-conversions-7.0.0.tgz#256b4e1882be7debbf01d05f0aa2039778ea080a" @@ -11850,6 +12178,11 @@ webpack-sources@^3.2.3: resolved "https://registry.yarnpkg.com/webpack-sources/-/webpack-sources-3.2.3.tgz#2d4daab8451fd4b240cc27055ff6a0c2ccea0cde" integrity sha512-/DyMEOrDgLKKIG0fmvtz+4dUX/3Ghozwgm6iPp8KRhvn+eQf9+Q7GWxVNMk3+uCPWfdXYC4ExGBckIXdFEfH1w== +webpack-virtual-modules@^0.5.0: + version "0.5.0" + resolved "https://registry.yarnpkg.com/webpack-virtual-modules/-/webpack-virtual-modules-0.5.0.tgz#362f14738a56dae107937ab98ea7062e8bdd3b6c" + integrity sha512-kyDivFZ7ZM0BVOUteVbDFhlRt7Ah/CSPwJdi8hBpkK7QLumUqdLtVfm/PX/hkcnrvr0i77fO5+TjZ94Pe+C9iw== + webpack-virtual-modules@^0.6.1: version "0.6.1" resolved "https://registry.yarnpkg.com/webpack-virtual-modules/-/webpack-virtual-modules-0.6.1.tgz#ac6fdb9c5adb8caecd82ec241c9631b7a3681b6f" @@ -11919,6 +12252,14 @@ whatwg-url@^11.0.0: tr46 "^3.0.0" webidl-conversions "^7.0.0" +whatwg-url@^5.0.0: + version "5.0.0" + resolved "https://registry.yarnpkg.com/whatwg-url/-/whatwg-url-5.0.0.tgz#966454e8765462e37644d3626f6742ce8b70965d" + integrity sha512-saE57nupxk6v3HY35+jzBwYa0rKSy0XR8JSxZPwgLr7ys0IBzhGviA1/TUGJLmSVqs8pb9AnvICXEuOHLprYTw== + dependencies: + tr46 "~0.0.3" + webidl-conversions "^3.0.0" + which-boxed-primitive@^1.0.2: version "1.0.2" resolved "https://registry.yarnpkg.com/which-boxed-primitive/-/which-boxed-primitive-1.0.2.tgz#13757bc89b209b049fe5d86430e21cf40a89a8e6" @@ -11976,7 +12317,7 @@ which@^1.3.1: dependencies: isexe "^2.0.0" -which@^2.0.1: +which@^2.0.1, which@^2.0.2: version "2.0.2" resolved "https://registry.yarnpkg.com/which/-/which-2.0.2.tgz#7c6a8dd0a636a0327e10b59c9286eee93f3f51b1" integrity sha512-BLI3Tl1TW3Pvl70l3yq3Y64i+awpwXqsGBYWkkqMtnbXgrMD+yj7rhW0kuEDxzJaYXGjEW5ogapKNMEKNMjibA==
b57fe214a2639a8ed0cce83e1f0a9ffcf44f3525
2023-04-21 22:24:06
Gilbert Szeto
fix(nodestore): retry on ServiceUnavailable exception (#47737)
false
retry on ServiceUnavailable exception (#47737)
fix
diff --git a/src/sentry/utils/kvstore/bigtable.py b/src/sentry/utils/kvstore/bigtable.py index aec73fb83a7509..13282726ff7b3c 100644 --- a/src/sentry/utils/kvstore/bigtable.py +++ b/src/sentry/utils/kvstore/bigtable.py @@ -172,11 +172,11 @@ def __decode_row(self, row: PartialRowData) -> Optional[bytes]: def set(self, key: str, value: bytes, ttl: Optional[timedelta] = None) -> None: try: return self._set(key, value, ttl) - except exceptions.InternalServerError: + except (exceptions.InternalServerError, exceptions.ServiceUnavailable): # Delete cached client before retry with self.__table_lock: del self.__table - # Retry once on InternalServerError + # Retry once on InternalServerError or ServiceUnavailable # 500 Received RST_STREAM with error code 2 # SENTRY-S6D return self._set(key, value, ttl)
2a852be875b36d4fe9bacd3e69f7729b0652f9f5
2022-10-17 21:58:18
Andrew Xue
feat(issue-alert): issue alert preview graph (#40043)
false
issue alert preview graph (#40043)
feat
diff --git a/static/app/views/alerts/create.spec.jsx b/static/app/views/alerts/create.spec.jsx index a4578c636602a8..ac21d3f7fbaa96 100644 --- a/static/app/views/alerts/create.spec.jsx +++ b/static/app/views/alerts/create.spec.jsx @@ -1,4 +1,5 @@ import selectEvent from 'react-select-event'; +import moment from 'moment'; import {initializeOrg} from 'sentry-test/initializeOrg'; import {act, render, screen, userEvent, waitFor} from 'sentry-test/reactTestingLibrary'; @@ -439,4 +440,58 @@ describe('ProjectAlertsCreate', function () { }); }); }); + + describe('test preview chart', () => { + const organization = TestStubs.Organization({features: ['issue-alert-preview']}); + afterEach(() => { + jest.clearAllMocks(); + }); + it('generate valid preview chart', async () => { + const mock = MockApiClient.addMockResponse({ + url: '/projects/org-slug/project-slug/rules/preview', + method: 'POST', + body: [ + {datetime: moment().subtract(2, 'days').format(), count: 1}, + {datetime: moment().subtract(1, 'days').format(), count: 2}, + {datetime: moment().format(), count: 3}, + ], + }); + createWrapper({organization}); + userEvent.click(screen.getByText('Generate Preview')); + await waitFor(() => { + expect(mock).toHaveBeenCalledWith( + expect.any(String), + expect.objectContaining({ + data: { + actionMatch: 'all', + conditions: [], + filterMatch: 'all', + filters: [], + frequency: 30, + }, + }) + ); + }); + expect(screen.getByText('Alerts Triggered')).toBeInTheDocument(); + expect(screen.getByText('Total Alerts')).toBeInTheDocument(); + }); + + it('invalid preview chart', async () => { + const mock = MockApiClient.addMockResponse({ + url: '/projects/org-slug/project-slug/rules/preview', + method: 'POST', + statusCode: 400, + }); + createWrapper({organization}); + userEvent.click(screen.getByText('Generate Preview')); + await waitFor(() => { + expect(mock).toHaveBeenCalled(); + }); + expect( + screen.getByText( + 'Previews are unavailable for this combination of conditions and filters' + ) + ).toBeInTheDocument(); + }); + }); }); diff --git a/static/app/views/alerts/rules/issue/index.tsx b/static/app/views/alerts/rules/issue/index.tsx index 2e7f6f3f44c2bc..9e3aeed931023e 100644 --- a/static/app/views/alerts/rules/issue/index.tsx +++ b/static/app/views/alerts/rules/issue/index.tsx @@ -15,6 +15,7 @@ import { } from 'sentry/actionCreators/indicator'; import {updateOnboardingTask} from 'sentry/actionCreators/onboardingTasks'; import Access from 'sentry/components/acl/access'; +import Feature from 'sentry/components/acl/feature'; import Alert from 'sentry/components/alert'; import Button from 'sentry/components/button'; import Confirm from 'sentry/components/confirm'; @@ -49,6 +50,7 @@ import { IssueAlertRuleAction, IssueAlertRuleActionTemplate, IssueAlertRuleConditionTemplate, + ProjectAlertRuleStats, UnsavedIssueAlertRule, } from 'sentry/types/alerts'; import {metric} from 'sentry/utils/analytics'; @@ -59,6 +61,7 @@ import recreateRoute from 'sentry/utils/recreateRoute'; import routeTitleGen from 'sentry/utils/routeTitle'; import withOrganization from 'sentry/utils/withOrganization'; import withProjects from 'sentry/utils/withProjects'; +import PreviewChart from 'sentry/views/alerts/rules/issue/previewChart'; import { CHANGE_ALERT_CONDITION_IDS, CHANGE_ALERT_PLACEHOLDERS_LABELS, @@ -134,7 +137,9 @@ type State = AsyncView['state'] & { [key: string]: string[]; }; environments: Environment[] | null; + previewError: null | string; project: Project; + ruleFireHistory: ProjectAlertRuleStats[] | null; uuid: null | string; duplicateTargetRule?: UnsavedIssueAlertRule | IssueAlertRule | null; ownership?: null | IssueOwnership; @@ -189,6 +194,7 @@ class IssueRuleEditor extends AsyncView<Props, State> { environments: [], uuid: null, project, + ruleFireHistory: null, }; const projectTeamIds = new Set(project.teams.map(({id}) => id)); @@ -301,6 +307,42 @@ class IssueRuleEditor extends AsyncView<Props, State> { } }; + fetchPreview = async () => { + const {organization} = this.props; + const {project, rule} = this.state; + + if (!rule) { + return; + } + this.setState({loadingPreview: true}); + try { + const response = await this.api.requestPromise( + `/projects/${organization.slug}/${project.slug}/rules/preview`, + { + method: 'POST', + data: { + conditions: rule?.conditions || [], + filters: rule?.filters || [], + actionMatch: rule?.actionMatch || 'all', + filterMatch: rule?.filterMatch || 'all', + frequency: rule?.frequency || 60, + }, + } + ); + this.setState({ + ruleFireHistory: response, + previewError: null, + loadingPreview: false, + }); + } catch (err) { + this.setState({ + previewError: + 'Previews are unavailable for this combination of conditions and filters', + loadingPreview: false, + }); + } + }; + fetchEnvironments() { const { params: {orgId}, @@ -759,6 +801,21 @@ class IssueRuleEditor extends AsyncView<Props, State> { ); } + renderPreviewGraph() { + const {ruleFireHistory, previewError} = this.state; + if (ruleFireHistory && !previewError) { + return <PreviewChart ruleFireHistory={ruleFireHistory} />; + } + if (previewError) { + return ( + <Alert type="error" showIcon> + {previewError} + </Alert> + ); + } + return null; + } + renderProjectSelect(disabled: boolean) { const {project: _selectedProject, projects, organization} = this.props; const {rule} = this.state; @@ -892,7 +949,7 @@ class IssueRuleEditor extends AsyncView<Props, State> { renderBody() { const {organization} = this.props; - const {project, rule, detailedError, loading, ownership} = this.state; + const {project, rule, detailedError, loading, ownership, loadingPreview} = this.state; const {actions, filters, conditions, frequency} = rule || {}; const environment = @@ -1148,6 +1205,28 @@ class IssueRuleEditor extends AsyncView<Props, State> { </StyledFieldHelp> </StyledListItem> {this.renderActionInterval(disabled)} + <Feature organization={organization} features={['issue-alert-preview']}> + <StyledListItem> + <StyledListItemSpaced> + <div> + {t('Preview history graph')} + <StyledFieldHelp> + {t( + 'Shows when this rule would have fired in the past 2 weeks' + )} + </StyledFieldHelp> + </div> + <Button + onClick={this.fetchPreview} + type="button" + disabled={loadingPreview} + > + Generate Preview + </Button> + </StyledListItemSpaced> + </StyledListItem> + {this.renderPreviewGraph()} + </Feature> <StyledListItem>{t('Establish ownership')}</StyledListItem> {this.renderRuleName(disabled)} {this.renderTeamSelect(disabled)} @@ -1182,6 +1261,11 @@ const StyledListItem = styled(ListItem)` font-size: ${p => p.theme.fontSizeExtraLarge}; `; +const StyledListItemSpaced = styled('div')` + display: flex; + justify-content: space-between; +`; + const StyledFieldHelp = styled(FieldHelp)` margin-top: 0; `; diff --git a/static/app/views/alerts/rules/issue/previewChart.tsx b/static/app/views/alerts/rules/issue/previewChart.tsx new file mode 100644 index 00000000000000..d0e8cbfe57016a --- /dev/null +++ b/static/app/views/alerts/rules/issue/previewChart.tsx @@ -0,0 +1,98 @@ +import styled from '@emotion/styled'; + +import {AreaChart, AreaChartSeries} from 'sentry/components/charts/areaChart'; +import {HeaderTitleLegend, SectionHeading} from 'sentry/components/charts/styles'; +import {Panel, PanelBody, PanelFooter} from 'sentry/components/panels'; +import Placeholder from 'sentry/components/placeholder'; +import {t} from 'sentry/locale'; +import space from 'sentry/styles/space'; +import {ProjectAlertRuleStats} from 'sentry/types/alerts'; +import getDynamicText from 'sentry/utils/getDynamicText'; + +type Props = { + ruleFireHistory: ProjectAlertRuleStats[]; +}; + +const PreviewChart = ({ruleFireHistory}: Props) => { + const renderChart = fireHistory => { + const series: AreaChartSeries = { + seriesName: 'Alerts Triggered', + data: fireHistory.map(alert => ({ + name: alert.date, + value: alert.count, + })), + emphasis: { + disabled: true, + }, + }; + + return ( + <AreaChart + isGroupedByDate + showTimeInTooltip + grid={{ + left: space(0.25), + right: space(2), + top: space(3), + bottom: 0, + }} + yAxis={{ + minInterval: 1, + }} + series={[series]} + /> + ); + }; + + const totalAlertsTriggered = ruleFireHistory.reduce((acc, curr) => acc + curr.count, 0); + + return ( + <Panel> + <StyledPanelBody withPadding> + <ChartHeader> + <HeaderTitleLegend>{t('Alerts Triggered')}</HeaderTitleLegend> + </ChartHeader> + {getDynamicText({ + value: renderChart(ruleFireHistory), + fixed: <Placeholder height="200px" testId="skeleton-ui" />, + })} + </StyledPanelBody> + <ChartFooter> + <FooterHeader>{t('Total Alerts')}</FooterHeader> + <FooterValue>{totalAlertsTriggered.toLocaleString()}</FooterValue> + </ChartFooter> + </Panel> + ); +}; + +export default PreviewChart; + +const ChartHeader = styled('div')` + margin-bottom: ${space(3)}; +`; + +const ChartFooter = styled(PanelFooter)` + display: flex; + align-items: center; + padding: ${space(1)} 20px; +`; + +const FooterHeader = styled(SectionHeading)` + display: flex; + align-items: center; + margin: 0; + font-weight: bold; + font-size: ${p => p.theme.fontSizeMedium}; + line-height: 1; +`; + +const FooterValue = styled('div')` + display: flex; + align-items: center; + margin: 0 ${space(1)}; +`; + +/* Override padding to make chart appear centered */ +const StyledPanelBody = styled(PanelBody)` + padding-right: ${space(0.75)}; +`;
e61ea4028c07b65d926b04a1dc8569b3b63c5ac5
2019-08-09 23:27:20
Dan Fuller
feat(api): Call snuba subscriptions endpoint to create alert rule subscriptions (SEN-827)
false
Call snuba subscriptions endpoint to create alert rule subscriptions (SEN-827)
feat
diff --git a/src/sentry/incidents/logic.py b/src/sentry/incidents/logic.py index b4356e3817e1c4..9e3b8c22a1e4e9 100644 --- a/src/sentry/incidents/logic.py +++ b/src/sentry/incidents/logic.py @@ -1,5 +1,6 @@ from __future__ import absolute_import +import uuid from collections import defaultdict from datetime import timedelta from uuid import uuid4 @@ -7,13 +8,16 @@ import pytz import six from dateutil.parser import parse as parse_date +from django.conf import settings from django.db import transaction from django.utils import timezone from sentry import analytics from sentry.api.event_search import get_snuba_query_args +from sentry.http import safe_urlopen from sentry.incidents.models import ( AlertRule, + AlertRuleAggregations, AlertRuleStatus, Incident, IncidentActivity, @@ -43,6 +47,10 @@ ) MAX_INITIAL_INCIDENT_PERIOD = timedelta(days=7) +alert_aggregation_to_snuba = { + AlertRuleAggregations.TOTAL: ('count()', '', 'count'), + AlertRuleAggregations.UNIQUE_USERS: ('uniq', 'tags[sentry:user]', 'unique_users'), +} class StatusAlreadyChangedError(Exception): @@ -624,6 +632,7 @@ def create_alert_rule( raise AlertRuleNameAlreadyUsedError() try: subscription_id = create_snuba_subscription( + project, dataset, query, aggregations, @@ -714,10 +723,13 @@ def update_alert_rule( old_subscription_id = alert_rule.subscription_id # If updating any details of the query, create a new subscription subscription_id = create_snuba_subscription( - alert_rule.dataset, - query, - aggregations, - time_window, + alert_rule.project, + SnubaDatasets(alert_rule.dataset), + query if query is not None else alert_rule.query, + aggregations if aggregations else [ + AlertRuleAggregations(agg) for agg in alert_rule.aggregations + ], + time_window if time_window else alert_rule.time_window, DEFAULT_ALERT_RULE_RESOLUTION, ) updated_fields['subscription_id'] = subscription_id @@ -730,10 +742,10 @@ def update_alert_rule( if subscription_id: delete_snuba_subscription(subscription_id) raise - finally: - if old_subscription_id: - # Once we're set up correctly, remove the previous subscription id. - delete_snuba_subscription(old_subscription_id) + + if old_subscription_id: + # Once we're set up correctly, remove the previous subscription id. + delete_snuba_subscription(old_subscription_id) return alert_rule @@ -766,15 +778,39 @@ def validate_alert_rule_query(query): get_snuba_query_args(query) -def create_snuba_subscription(dataset, query, aggregations, time_window, resolution): +def create_snuba_subscription(project, dataset, query, aggregations, time_window, resolution): """ Creates a subscription to a snuba query. - :param alert_rule: The alert rule to create the subscription for + :param project: The project we're applying the query to + :param dataset: The snuba dataset to query and aggregate over + :param query: An event search query that we can parse and convert into a + set of Snuba conditions + :param aggregations: A list of aggregations to calculate over the time + window + :param time_window: The time window to aggregate over + :param resolution: How often to receive updates/bucket size :return: A uuid representing the subscription id. """ - # TODO: Implement - return uuid4() + # TODO: Might make sense to move this into snuba if we have wider use for + # it. + resp = safe_urlopen( + settings.SENTRY_SNUBA + '/subscriptions', + 'POST', + json={ + 'project_id': project.id, + 'dataset': dataset.value, + # We only care about conditions here. Filter keys only matter for + # filtering to project and groups. Projects are handled with an + # explicit param, and groups can't be queried here. + 'conditions': get_snuba_query_args(query)['conditions'], + 'aggregates': [alert_aggregation_to_snuba[agg] for agg in aggregations], + 'time_window': time_window, + 'resolution': resolution, + }, + ) + resp.raise_for_status() + return uuid.UUID(resp.json()['subscription_id']) def delete_snuba_subscription(subscription_id): diff --git a/tests/sentry/net/test_socket.py b/tests/sentry/net/test_socket.py index ae550bc14ac15f..e405eafbe0ced4 100644 --- a/tests/sentry/net/test_socket.py +++ b/tests/sentry/net/test_socket.py @@ -17,8 +17,11 @@ class SocketTest(TestCase): @override_blacklist('10.0.0.0/8', '127.0.0.1') def test_is_ipaddress_allowed(self): + is_ipaddress_allowed.cache_clear() assert is_ipaddress_allowed('127.0.0.1') is False + is_ipaddress_allowed.cache_clear() assert is_ipaddress_allowed('10.0.1.1') is False + is_ipaddress_allowed.cache_clear() assert is_ipaddress_allowed('1.1.1.1') is True @override_blacklist('10.0.0.0/8', '127.0.0.1')
796c91bc820965677afdd58870ae2d3976d5be92
2018-11-21 17:25:18
Markus Unterwaditzer
fix: Only fork when running rust code (#10711)
false
Only fork when running rust code (#10711)
fix
diff --git a/src/sentry/management/commands/serve_normalize.py b/src/sentry/management/commands/serve_normalize.py index 802078bd978242..40d061d31297fb 100644 --- a/src/sentry/management/commands/serve_normalize.py +++ b/src/sentry/management/commands/serve_normalize.py @@ -22,6 +22,35 @@ from django.utils.encoding import force_str +def catch_errors(f): + def wrapper(*args, **kwargs): + error = None + try: + return f(*args, **kwargs) + except Exception as e: + error = force_str(e.message) + ' ' + force_str(traceback.format_exc()) + + try: + return encode({ + 'result': None, + 'error': error, + 'metrics': None + }) + except (ValueError, TypeError) as e: + try: + # Encoding error, try to send the exception instead + return encode({ + 'result': None, + 'error': force_str(e.message) + ' ' + force_str(traceback.format_exc()), + 'metrics': None, + 'encoding_error': True, + }) + except Exception: + return b'{}' + + return wrapper + + # Here's where the normalization itself happens def process_event(data, meta): from sentry.event_manager import EventManager, get_hashes_for_event @@ -59,52 +88,24 @@ def encode(data): return json.dumps(data) -def handle_data(pipe, data): - @catch_errors - def inner(data): - mc = MetricCollector() +@catch_errors +def handle_data(data): + mc = MetricCollector() - metrics_before = mc.collect_metrics() - data, meta = decode(data) - rv = process_event(data, meta) - metrics_after = mc.collect_metrics() + metrics_before = mc.collect_metrics() + data, meta = decode(data) + rv = process_event(data, meta) + metrics_after = mc.collect_metrics() - return encode({ - 'result': rv, - 'metrics': {'before': metrics_before, 'after': metrics_after}, - 'error': None - }) + return encode({ + 'result': rv, + 'metrics': {'before': metrics_before, 'after': metrics_after}, + 'error': None + }) - pipe.send(inner(data)) - -def catch_errors(f): - def wrapper(*args, **kwargs): - error = None - try: - return f(*args, **kwargs) - except Exception as e: - error = force_str(e.message) + ' ' + force_str(traceback.format_exc()) - - try: - return encode({ - 'result': None, - 'error': error, - 'metrics': None - }) - except (ValueError, TypeError) as e: - try: - # Encoding error, try to send the exception instead - return encode({ - 'result': None, - 'error': force_str(e.message) + ' ' + force_str(traceback.format_exc()), - 'metrics': None, - 'encoding_error': True, - }) - except Exception: - return b'{}' - - return wrapper +def handle_data_piped(pipe, data): + pipe.send(handle_data(data)) class MetricCollector(object): @@ -162,10 +163,18 @@ def handle(self): self.request.close() def handle_data(self): + from sentry.event_manager import ENABLE_RUST + if not ENABLE_RUST: + return handle_data(self.data) + @catch_errors def inner(): + # TODO: Remove this contraption once we no longer get segfaults parent_conn, child_conn = multiprocessing.Pipe() - p = multiprocessing.Process(target=handle_data, args=(child_conn, self.data,)) + p = multiprocessing.Process( + target=handle_data_piped, + args=(child_conn, self.data,) + ) p.start() p.join(1) assert parent_conn.poll(), "Process crashed"
e21c9ad6416c5e1f6edb0ebc14eb95eba17cf05e
2023-07-18 02:16:44
Armen Zambrano G
ref(event_manager): Fix typing issues for event_manager (#52974)
false
Fix typing issues for event_manager (#52974)
ref
diff --git a/pyproject.toml b/pyproject.toml index bfae3de90ebebe..17aba86dc76867 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -442,7 +442,6 @@ module = [ "sentry.discover.endpoints.discover_key_transactions", "sentry.discover.endpoints.serializers", "sentry.discover.tasks", - "sentry.event_manager", "sentry.eventstore.base", "sentry.eventstore.compressor", "sentry.eventstore.models", @@ -1156,7 +1155,6 @@ module = [ "tests.sentry.eventstore.test_models", "tests.sentry.eventstream.kafka.test_protocol", "tests.sentry.eventstream.test_eventstream", - "tests.sentry.eventtypes.test_default", "tests.sentry.eventtypes.test_error", "tests.sentry.features.test_manager", "tests.sentry.grouping.test_enhancer", diff --git a/src/sentry/attachments/__init__.py b/src/sentry/attachments/__init__.py index 64ad58c27512e4..e8ebbcc3cd055a 100644 --- a/src/sentry/attachments/__init__.py +++ b/src/sentry/attachments/__init__.py @@ -1,4 +1,4 @@ -__all__ = ["attachment_cache", "CachedAttachment"] +__all__ = ["attachment_cache", "CachedAttachment", "MissingAttachmentChunks"] from django.conf import settings diff --git a/src/sentry/event_manager.py b/src/sentry/event_manager.py index 67d7f198d16f63..b55272f355daf4 100644 --- a/src/sentry/event_manager.py +++ b/src/sentry/event_manager.py @@ -19,7 +19,6 @@ Optional, Sequence, Tuple, - Type, TypedDict, Union, cast, @@ -55,15 +54,8 @@ from sentry.culprit import generate_culprit from sentry.dynamic_sampling import LatestReleaseBias, LatestReleaseParams from sentry.eventstore.processing import event_processing_store -from sentry.eventtypes import ( - CspEvent, - DefaultEvent, - ErrorEvent, - ExpectCTEvent, - ExpectStapleEvent, - HpkpEvent, - TransactionEvent, -) +from sentry.eventtypes import EventType +from sentry.eventtypes.transaction import TransactionEvent from sentry.grouping.api import ( BackgroundGroupingConfigLoader, GroupingConfig, @@ -660,7 +652,7 @@ def save_error_events( with metrics.timer("event_manager.save_attachments"): save_attachments(cache_key, attachments, job) - metric_tags = {"from_relay": "_relay_processed" in job["data"]} + metric_tags = {"from_relay": str("_relay_processed" in job["data"])} metrics.timing( "events.latency", @@ -1260,13 +1252,15 @@ def _tsdb_record_all_metrics(jobs: Sequence[Job]) -> None: records.append((TSDBModel.users_affected_by_project, project_id, (user.tag_value,))) if incrs: - tsdb.incr_multi(incrs, timestamp=event.datetime, environment_id=environment.id) + tsdb.backend.incr_multi(incrs, timestamp=event.datetime, environment_id=environment.id) if records: - tsdb.record_multi(records, timestamp=event.datetime, environment_id=environment.id) + tsdb.backend.record_multi( + records, timestamp=event.datetime, environment_id=environment.id + ) if frequencies: - tsdb.record_frequency_multi(frequencies, timestamp=event.datetime) + tsdb.backend.record_frequency_multi(frequencies, timestamp=event.datetime) @metrics.wraps("save_event.nodestore_save_many") @@ -1438,17 +1432,6 @@ def _get_event_user_impl( return euser -EventType = Union[ - DefaultEvent, - ErrorEvent, - CspEvent, - HpkpEvent, - ExpectCTEvent, - ExpectStapleEvent, - TransactionEvent, -] - - def get_event_type(data: Mapping[str, Any]) -> EventType: return eventtypes.get(data.get("type", "default"))() @@ -1922,7 +1905,7 @@ def _process_existing_aggregate( return bool(is_regression) -Attachment = Type[CachedAttachment] +Attachment = CachedAttachment def discard_event(job: Job, attachments: Sequence[Attachment]) -> None: @@ -1938,7 +1921,7 @@ def discard_event(job: Job, attachments: Sequence[Attachment]) -> None: project = job["event"].project - quotas.refund( + quotas.backend.refund( project, key=job["project_key"], timestamp=job["start_time"], @@ -1975,7 +1958,7 @@ def discard_event(job: Job, attachments: Sequence[Attachment]) -> None: ) if attachment_quantity: - quotas.refund( + quotas.backend.refund( project, key=job["project_key"], timestamp=job["start_time"], @@ -2099,7 +2082,7 @@ def filter_attachments_for_group(attachments: list[Attachment], job: Job) -> lis cache.set(crashreports_key, max_crashreports, CRASH_REPORT_TIMEOUT) if refund_quantity: - quotas.refund( + quotas.backend.refund( project, key=job["project_key"], timestamp=job["start_time"], diff --git a/src/sentry/eventtypes/__init__.py b/src/sentry/eventtypes/__init__.py index 143f2e28eb93f2..c370aa787f3da2 100644 --- a/src/sentry/eventtypes/__init__.py +++ b/src/sentry/eventtypes/__init__.py @@ -1,3 +1,5 @@ +from typing import Union + from .base import DefaultEvent from .error import ErrorEvent from .generic import GenericEvent @@ -17,3 +19,13 @@ get = default_manager.get register = default_manager.register + +EventType = Union[ + DefaultEvent, + ErrorEvent, + CspEvent, + HpkpEvent, + ExpectCTEvent, + ExpectStapleEvent, + TransactionEvent, +] diff --git a/src/sentry/grouping/result.py b/src/sentry/grouping/result.py index 38fcfafac5a0f9..976545c46d2eb5 100644 --- a/src/sentry/grouping/result.py +++ b/src/sentry/grouping/result.py @@ -1,9 +1,9 @@ from dataclasses import dataclass from typing import Any, Dict, List, Optional, Sequence, TypedDict, Union +from sentry.db.models import NodeData from sentry.utils.safe import get_path, safe_execute, set_path -EventData = Dict[str, Any] EventMetadata = Dict[str, Any] @@ -63,7 +63,7 @@ def _strip_tree_label(tree_label: TreeLabel, truncate: bool = False) -> Stripped return rv -def _write_tree_labels(tree_labels: Sequence[Optional[TreeLabel]], event_data: EventData) -> None: +def _write_tree_labels(tree_labels: Sequence[Optional[TreeLabel]], event_data: NodeData) -> None: event_labels: List[Optional[StrippedTreeLabel]] = [] event_data["hierarchical_tree_labels"] = event_labels @@ -97,7 +97,7 @@ class CalculatedHashes: hierarchical_hashes: Sequence[str] tree_labels: Sequence[Optional[TreeLabel]] - def write_to_event(self, event_data: EventData) -> None: + def write_to_event(self, event_data: NodeData) -> None: event_data["hashes"] = self.hashes if self.hierarchical_hashes: @@ -106,7 +106,7 @@ def write_to_event(self, event_data: EventData) -> None: safe_execute(_write_tree_labels, self.tree_labels, event_data, _with_transaction=False) @classmethod - def from_event(cls, event_data: EventData) -> Optional["CalculatedHashes"]: + def from_event(cls, event_data: NodeData) -> Optional["CalculatedHashes"]: hashes = event_data.get("hashes") hierarchical_hashes = event_data.get("hierarchical_hashes") or [] tree_labels = event_data.get("hierarchical_tree_labels") or [] diff --git a/src/sentry/tsdb/__init__.py b/src/sentry/tsdb/__init__.py index 1aa88203fec695..847cf864aa0f7a 100644 --- a/src/sentry/tsdb/__init__.py +++ b/src/sentry/tsdb/__init__.py @@ -5,6 +5,7 @@ from .base import BaseTSDB from .dummy import DummyTSDB -LazyServiceWrapper( +backend = LazyServiceWrapper( BaseTSDB, settings.SENTRY_TSDB, settings.SENTRY_TSDB_OPTIONS, dangerous=[DummyTSDB] -).expose(locals()) +) +backend.expose(locals()) diff --git a/tests/sentry/eventtypes/test_default.py b/tests/sentry/eventtypes/test_default.py index 5c9443de01af85..de36b5c3820f72 100644 --- a/tests/sentry/eventtypes/test_default.py +++ b/tests/sentry/eventtypes/test_default.py @@ -1,4 +1,4 @@ -from sentry.eventtypes import DefaultEvent +from sentry.eventtypes.base import DefaultEvent from sentry.testutils import TestCase from sentry.testutils.silo import region_silo_test
4722b205f33ca5d5041e9a8a996ef177358f9c1c
2023-01-23 17:52:36
Riccardo Busetti
fix(metrics): Fix project id meta type transformation [TET-645] (#43336)
false
Fix project id meta type transformation [TET-645] (#43336)
fix
diff --git a/src/sentry/snuba/metrics/query_builder.py b/src/sentry/snuba/metrics/query_builder.py index ab252415ee636c..52c6d9959c322d 100644 --- a/src/sentry/snuba/metrics/query_builder.py +++ b/src/sentry/snuba/metrics/query_builder.py @@ -10,6 +10,7 @@ ) from datetime import datetime, timedelta +from enum import Enum from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union from snuba_sdk import ( @@ -473,6 +474,60 @@ def get_metric_object_from_metric_field( return metric_object_factory(op=metric_field.op, metric_mri=metric_field.metric_mri) +class AliasMetaType(Enum): + """ + This class represents the type of an alias returned from a SNQL query. + + This type does not refer to the result type but rather the type of element to which the alias refers in the query + (e.g., MetricField, MetricGroupByField, dataset real columns, or simple tag). + """ + + TAG = 0 + DATASET_COLUMN = 1 + TIME_COLUMN = 2 + SELECT_METRIC_FIELD = 3 + GROUP_BY_METRIC_FIELD = 4 + + +def get_alias_meta_type( + returned_alias: str, + alias_to_metric_group_by_field: Dict[str, MetricGroupByField], +) -> Tuple[AliasMetaType, Tuple[Optional[str], str]]: + # This logic is a rewrite of the logic below, which was convoluted and not very expressive. + # + # Column name could be either a mri, ["bucketed_time"] or a tag or a dataset col like + # "project_id" or "metric_id". + # + # is_tag = parsed_alias in alias_to_metric_group_by_field.keys() + # is_time_col = parsed_alias in [TS_COL_GROUP] + # is_dataset_col = parsed_alias in DATASET_COLUMNS + parsed_expr = parse_expression(returned_alias) + parsed_op, parsed_alias = parsed_expr + + if parsed_alias in alias_to_metric_group_by_field: + field = alias_to_metric_group_by_field[parsed_alias].field + # This specific check is performed to extract the field from the MetricGroupByField in order + # to understand the target column. + if isinstance(field, str): + if field in DATASET_COLUMNS: + return AliasMetaType.DATASET_COLUMN, parsed_expr + elif field in [TS_COL_GROUP]: + return AliasMetaType.TIME_COLUMN, parsed_expr + else: + return AliasMetaType.TAG, parsed_expr + elif isinstance(field, MetricField): + return AliasMetaType.GROUP_BY_METRIC_FIELD, parsed_expr + + # This logic has been copied from the logic before that handles specific edge cases that were assumed by the author + # of the original function. + if parsed_alias in DATASET_COLUMNS: + return AliasMetaType.DATASET_COLUMN, parsed_expr + elif parsed_alias in [TS_COL_GROUP]: + return AliasMetaType.TIME_COLUMN, parsed_expr + + return AliasMetaType.SELECT_METRIC_FIELD, parsed_expr + + def translate_meta_results( meta: Sequence[Dict[str, str]], alias_to_metric_field: Dict[str, MetricField], @@ -487,19 +542,13 @@ def translate_meta_results( """ results = [] for record in meta: - operation, column_name = parse_expression(record["name"]) - - # Column name could be either a mri, ["bucketed_time"] or a tag or a dataset col like - # "project_id" or "metric_id" - is_tag = column_name in alias_to_metric_group_by_field.keys() - is_time_col = column_name in [TS_COL_GROUP] - is_dataset_col = column_name in DATASET_COLUMNS + alias_type, (parsed_op, parsed_alias) = get_alias_meta_type( + record["name"], alias_to_metric_group_by_field + ) - if not (is_tag or is_time_col or is_dataset_col): - # This handles two cases where we have an expression with an operation and an mri, - # or a derived metric mri that has no associated operation + if alias_type == AliasMetaType.SELECT_METRIC_FIELD: try: - record["name"] = get_operation_with_public_name(operation, column_name) + record["name"] = get_operation_with_public_name(parsed_op, parsed_alias) if COMPOSITE_ENTITY_CONSTITUENT_ALIAS in record["name"]: # Since instances of CompositeEntityDerivedMetric will not have meta data as they are computed post # query, it suffices to set the type of that composite derived metric to any of the types of its @@ -543,26 +592,21 @@ def translate_meta_results( # For example, If we have two constituents of types "UInt64" and "Float64", # then there inferred type would be "Float64" continue - else: - if is_tag: - # since we changed value from int to str we need - # also want to change type - metric_groupby_field = alias_to_metric_group_by_field[record["name"]] - if isinstance(metric_groupby_field.field, MetricField): - defined_parent_meta_type = get_metric_object_from_metric_field( - metric_groupby_field.field - ).get_meta_type() - else: - defined_parent_meta_type = None - - record["type"] = ( - "string" if defined_parent_meta_type is None else defined_parent_meta_type - ) - elif is_time_col or is_dataset_col: - record["name"] = column_name + elif alias_type == AliasMetaType.GROUP_BY_METRIC_FIELD: + metric_groupby_field = alias_to_metric_group_by_field[record["name"]] + defined_parent_meta_type = get_metric_object_from_metric_field( + metric_groupby_field.field + ).get_meta_type() + + record["type"] = defined_parent_meta_type + elif alias_type == AliasMetaType.TAG: + record["type"] = "string" + elif alias_type == AliasMetaType.DATASET_COLUMN or alias_type == AliasMetaType.TIME_COLUMN: + record["name"] = parsed_alias if record not in results: results.append(record) + return sorted(results, key=lambda elem: elem["name"]) diff --git a/tests/sentry/snuba/metrics/test_metrics_layer/test_metrics_enhanced_performance.py b/tests/sentry/snuba/metrics/test_metrics_layer/test_metrics_enhanced_performance.py index ac84674ddd0e27..c589534ac7338a 100644 --- a/tests/sentry/snuba/metrics/test_metrics_layer/test_metrics_enhanced_performance.py +++ b/tests/sentry/snuba/metrics/test_metrics_layer/test_metrics_enhanced_performance.py @@ -515,7 +515,7 @@ def test_query_with_order_by_valid_str_field(self): assert data["meta"] == sorted( [ {"name": "count(transaction.duration)", "type": "UInt64"}, - {"name": "project_id", "type": "string"}, + {"name": "project_id", "type": "UInt64"}, ], key=lambda elem: elem["name"], ) @@ -1069,7 +1069,7 @@ def test_groupby_aliasing_with_multiple_groups_and_orderby(self): {"name": "bucketed_time", "type": "DateTime('Universal')"}, {"name": "p50_fcp", "type": "Float64"}, {"name": "p50_lcp", "type": "Float64"}, - {"name": "project", "type": "string"}, + {"name": "project", "type": "UInt64"}, {"name": "project_alias", "type": "string"}, {"name": "transaction_group", "type": "string"}, ], diff --git a/tests/sentry/snuba/metrics/test_query_builder.py b/tests/sentry/snuba/metrics/test_query_builder.py index 1c1dcee769072c..a50b23adc5604c 100644 --- a/tests/sentry/snuba/metrics/test_query_builder.py +++ b/tests/sentry/snuba/metrics/test_query_builder.py @@ -1051,6 +1051,9 @@ def test_translate_meta_results(): {"name": "transaction", "type": "UInt64"}, {"name": "project_id", "type": "UInt64"}, {"name": "metric_id", "type": "UInt64"}, + {"name": "bucketed_time", "type": "UInt64"}, + {"name": "project.id", "type": "UInt64"}, + {"name": "time", "type": "UInt64"}, ] assert translate_meta_results( meta, @@ -1070,6 +1073,8 @@ def test_translate_meta_results(): alias="team_key_transaction", ) ), + "project.id": MetricGroupByField(field="project_id"), + "time": MetricGroupByField(field="bucketed_time"), }, ) == sorted( [ @@ -1078,6 +1083,9 @@ def test_translate_meta_results(): {"name": "transaction", "type": "string"}, {"name": "project_id", "type": "UInt64"}, {"name": "metric_id", "type": "UInt64"}, + {"name": "bucketed_time", "type": "UInt64"}, + {"name": "project.id", "type": "UInt64"}, + {"name": "time", "type": "UInt64"}, ], key=lambda elem: elem["name"], ) diff --git a/tests/snuba/api/endpoints/test_organization_events_mep.py b/tests/snuba/api/endpoints/test_organization_events_mep.py index 78c6c570e330e1..eb1fb1900b1062 100644 --- a/tests/snuba/api/endpoints/test_organization_events_mep.py +++ b/tests/snuba/api/endpoints/test_organization_events_mep.py @@ -2145,11 +2145,3 @@ def test_custom_measurement_size_filtering(self): @pytest.mark.xfail(reason="Having not supported") def test_having_condition(self): super().test_having_condition() - - @pytest.mark.xfail(reason="Meta for project_id is wrong") - def test_project_id(self): - super().test_project_id() - - @pytest.mark.xfail(reason="Meta for project.id is wrong") - def test_project_dot_id(self): - super().test_project_dot_id()
5e93e718d638c638de95be3cc10bc786c31f393e
2020-01-15 00:03:21
Stephen Cefali
feat(ui): adds callback for AsyncComponent when all endpoints are loaded (#16427)
false
adds callback for AsyncComponent when all endpoints are loaded (#16427)
feat
diff --git a/src/sentry/static/sentry/app/components/asyncComponent.tsx b/src/sentry/static/sentry/app/components/asyncComponent.tsx index 66bf71c7899718..123e9cd0bf408f 100644 --- a/src/sentry/static/sentry/app/components/asyncComponent.tsx +++ b/src/sentry/static/sentry/app/components/asyncComponent.tsx @@ -292,23 +292,35 @@ export default class AsyncComponent< // Allow children to implement this } - handleRequestSuccess({stateKey, data, jqXHR}, initialRequest?: boolean) { - this.setState(prevState => { - const state = { - [stateKey]: data, - // TODO(billy): This currently fails if this request is retried by SudoModal - [`${stateKey}PageLinks`]: jqXHR && jqXHR.getResponseHeader('Link'), - }; + onLoadAllEndpointsSuccess() { + // Allow children to implement this + } - if (initialRequest) { - state.remainingRequests = prevState.remainingRequests! - 1; - state.loading = prevState.remainingRequests! > 1; - state.reloading = prevState.reloading && state.loading; - this.markShouldMeasure({remainingRequests: state.remainingRequests}); + handleRequestSuccess({stateKey, data, jqXHR}, initialRequest?: boolean) { + this.setState( + prevState => { + const state = { + [stateKey]: data, + // TODO(billy): This currently fails if this request is retried by SudoModal + [`${stateKey}PageLinks`]: jqXHR && jqXHR.getResponseHeader('Link'), + }; + + if (initialRequest) { + state.remainingRequests = prevState.remainingRequests! - 1; + state.loading = prevState.remainingRequests! > 1; + state.reloading = prevState.reloading && state.loading; + this.markShouldMeasure({remainingRequests: state.remainingRequests}); + } + + return state; + }, + () => { + //if everything is loaded and we don't have an error, call the callback + if (this.state.remainingRequests === 0 && !this.state.error) { + this.onLoadAllEndpointsSuccess(); + } } - - return state; - }); + ); this.onRequestSuccess({stateKey, data, jqXHR}); } diff --git a/tests/js/spec/components/asyncComponent.spec.jsx b/tests/js/spec/components/asyncComponent.spec.jsx index a8324b594e91ce..91c83dbf471234 100644 --- a/tests/js/spec/components/asyncComponent.spec.jsx +++ b/tests/js/spec/components/asyncComponent.spec.jsx @@ -55,4 +55,51 @@ describe('AsyncComponent', function() { .text() ).toEqual('oops there was a problem'); }); + + describe('multi-route component', () => { + class MultiRouteComponent extends TestAsyncComponent { + getEndpoints() { + return [['data', '/some/path/to/something/'], ['project', '/another/path/here']]; + } + } + + it('calls onLoadAllEndpointsSuccess when all endpoints have been loaded', () => { + jest.useFakeTimers(); + jest.spyOn(Client.prototype, 'request').mockImplementation((url, options) => { + const timeout = url.includes('something') ? 100 : 50; + setTimeout( + () => + options.success({ + message: 'good', + }), + timeout + ); + }); + const mockOnAllEndpointsSuccess = jest.spyOn( + MultiRouteComponent.prototype, + 'onLoadAllEndpointsSuccess' + ); + + const wrapper = shallow(<MultiRouteComponent />); + + expect(wrapper.state('loading')).toEqual(true); + expect(wrapper.state('remainingRequests')).toEqual(2); + + jest.advanceTimersByTime(40); + expect(wrapper.state('loading')).toEqual(true); + expect(wrapper.state('remainingRequests')).toEqual(2); + + jest.advanceTimersByTime(40); + expect(wrapper.state('loading')).toEqual(true); + expect(wrapper.state('remainingRequests')).toEqual(1); + expect(mockOnAllEndpointsSuccess).not.toHaveBeenCalled(); + + jest.advanceTimersByTime(40); + expect(wrapper.state('loading')).toEqual(false); + expect(wrapper.state('remainingRequests')).toEqual(0); + expect(mockOnAllEndpointsSuccess).toHaveBeenCalled(); + + jest.restoreAllMocks(); + }); + }); });
b5b91d9daef76a1975a21d17ce956d900a5bfd62
2023-07-25 23:53:50
Malachi Willey
feat(most-helpful-event): Add event dropdown UI (#53377)
false
Add event dropdown UI (#53377)
feat
diff --git a/static/app/views/issueDetails/groupDetails.tsx b/static/app/views/issueDetails/groupDetails.tsx index f57c8e9a9b9ece..3d5bce53874099 100644 --- a/static/app/views/issueDetails/groupDetails.tsx +++ b/static/app/views/issueDetails/groupDetails.tsx @@ -248,14 +248,15 @@ function useEventApiQuery({ const hasMostHelpfulEventFeature = organization.features.includes( 'issue-details-most-helpful-event' ); - const eventIdUrl = eventId ?? (hasMostHelpfulEventFeature ? 'helpful' : 'latest'); + const eventIdUrl = eventId ?? (hasMostHelpfulEventFeature ? 'recommended' : 'latest'); const helpfulEventQuery = hasMostHelpfulEventFeature && typeof location.query.query === 'string' ? location.query.query : undefined; + const endpointEventId = eventIdUrl === 'recommended' ? 'helpful' : eventIdUrl; const queryKey: ApiQueryKey = [ - `/issues/${groupId}/events/${eventIdUrl}/`, + `/issues/${groupId}/events/${endpointEventId}/`, { query: getGroupEventDetailsQueryData({ environments, @@ -267,7 +268,7 @@ function useEventApiQuery({ const tab = getCurrentTab({router}); const isOnDetailsTab = tab === Tab.DETAILS; - const isLatestOrHelpfulEvent = eventIdUrl === 'latest' || eventIdUrl === 'helpful'; + const isLatestOrHelpfulEvent = eventIdUrl === 'latest' || eventIdUrl === 'recommended'; const latestOrHelpfulEvent = useApiQuery<Event>(queryKey, { // Latest/helpful event will change over time, so only cache for 30 seconds staleTime: 30000, diff --git a/static/app/views/issueDetails/groupEventCarousel.spec.tsx b/static/app/views/issueDetails/groupEventCarousel.spec.tsx index c8183f68094e7b..a723ee174f5dad 100644 --- a/static/app/views/issueDetails/groupEventCarousel.spec.tsx +++ b/static/app/views/issueDetails/groupEventCarousel.spec.tsx @@ -1,5 +1,8 @@ +import {browserHistory} from 'react-router'; + import {render, screen, userEvent, within} from 'sentry-test/reactTestingLibrary'; +import * as useMedia from 'sentry/utils/useMedia'; import {GroupEventCarousel} from 'sentry/views/issueDetails/groupEventCarousel'; describe('GroupEventCarousel', () => { @@ -21,7 +24,7 @@ describe('GroupEventCarousel', () => { }; beforeEach(() => { - jest.clearAllMocks(); + jest.restoreAllMocks(); Object.assign(navigator, { clipboard: { writeText: jest.fn().mockResolvedValue(''), @@ -30,13 +33,47 @@ describe('GroupEventCarousel', () => { window.open = jest.fn(); }); + it('can use event dropdown to navigate events', async () => { + // Because it isn't rendered on smaller screens + jest.spyOn(useMedia, 'default').mockReturnValue(true); + + render(<GroupEventCarousel {...defaultProps} />, { + organization: TestStubs.Organization({ + features: [ + 'issue-details-most-helpful-event', + 'issue-details-most-helpful-event-ui', + ], + }), + }); + + await userEvent.click(screen.getByRole('button', {name: /recommended event/i})); + await userEvent.click(screen.getByRole('option', {name: /oldest event/i})); + + expect(browserHistory.push).toHaveBeenCalledWith({ + pathname: '/organizations/org-slug/issues/group-id/events/oldest/', + query: {referrer: 'oldest-event'}, + }); + + await userEvent.click(screen.getByRole('button', {name: /oldest event/i})); + await userEvent.click(screen.getByRole('option', {name: /latest event/i})); + + expect(browserHistory.push).toHaveBeenCalledWith({ + pathname: '/organizations/org-slug/issues/group-id/events/oldest/', + query: {referrer: 'oldest-event'}, + }); + + await userEvent.click(screen.getByRole('button', {name: /latest event/i})); + await userEvent.click(screen.getByRole('option', {name: /recommended event/i})); + + expect(browserHistory.push).toHaveBeenCalledWith({ + pathname: '/organizations/org-slug/issues/group-id/events/recommended/', + query: {referrer: 'recommended-event'}, + }); + }); + it('can navigate next/previous events', () => { render(<GroupEventCarousel {...defaultProps} />); - expect(screen.getByLabelText(/First Event/)).toHaveAttribute( - 'href', - `/organizations/org-slug/issues/group-id/events/oldest/?referrer=oldest-event` - ); expect(screen.getByLabelText(/Previous Event/)).toHaveAttribute( 'href', `/organizations/org-slug/issues/group-id/events/prev-event-id/?referrer=previous-event` @@ -45,10 +82,6 @@ describe('GroupEventCarousel', () => { 'href', `/organizations/org-slug/issues/group-id/events/next-event-id/?referrer=next-event` ); - expect(screen.getByLabelText(/Latest Event/)).toHaveAttribute( - 'href', - `/organizations/org-slug/issues/group-id/events/latest/?referrer=latest-event` - ); }); it('can copy event ID', async () => { diff --git a/static/app/views/issueDetails/groupEventCarousel.tsx b/static/app/views/issueDetails/groupEventCarousel.tsx index a9ae42b655d67b..24e5af0e101adc 100644 --- a/static/app/views/issueDetails/groupEventCarousel.tsx +++ b/static/app/views/issueDetails/groupEventCarousel.tsx @@ -1,9 +1,11 @@ +import {browserHistory} from 'react-router'; import {useTheme} from '@emotion/react'; import styled from '@emotion/styled'; import moment from 'moment-timezone'; import {addSuccessMessage} from 'sentry/actionCreators/indicator'; import {Button, ButtonProps} from 'sentry/components/button'; +import {CompactSelect} from 'sentry/components/compactSelect'; import DateTime from 'sentry/components/dateTime'; import {DropdownMenu} from 'sentry/components/dropdownMenu'; import {Tooltip} from 'sentry/components/tooltip'; @@ -32,6 +34,7 @@ import useCopyToClipboard from 'sentry/utils/useCopyToClipboard'; import {useLocation} from 'sentry/utils/useLocation'; import useMedia from 'sentry/utils/useMedia'; import useOrganization from 'sentry/utils/useOrganization'; +import {useParams} from 'sentry/utils/useParams'; import {normalizeUrl} from 'sentry/utils/withDomainRequired'; import EventCreatedTooltip from 'sentry/views/issueDetails/eventCreatedTooltip'; @@ -52,9 +55,23 @@ type EventNavigationButtonProps = { eventId?: string | null; }; +enum EventNavDropdownOption { + RECOMMENDED = 'recommended', + LATEST = 'latest', + OLDEST = 'oldest', + ALL = 'all', +} + const BUTTON_SIZE = 'sm'; const BUTTON_ICON_SIZE = 'sm'; +const EVENT_NAV_DROPDOWN_OPTIONS = [ + {value: EventNavDropdownOption.RECOMMENDED, label: 'Recommended Event'}, + {value: EventNavDropdownOption.LATEST, label: 'Latest Event'}, + {value: EventNavDropdownOption.OLDEST, label: 'Oldest Event'}, + {options: [{value: EventNavDropdownOption.ALL, label: 'View All Events'}]}, +]; + const copyToClipboard = (value: string) => { navigator.clipboard .writeText(value) @@ -106,11 +123,74 @@ function EventNavigationButton({ ); } +function EventNavigationDropdown({group}: {group: Group}) { + const location = useLocation(); + const params = useParams<{eventId?: string}>(); + const theme = useTheme(); + const organization = useOrganization(); + const largeViewport = useMedia(`(min-width: ${theme.breakpoints.large})`); + + const isHelpfulEventUiEnabled = + organization.features.includes('issue-details-most-helpful-event') && + organization.features.includes('issue-details-most-helpful-event-ui'); + + if (!isHelpfulEventUiEnabled || !largeViewport) { + return null; + } + + const getSelectedOption = () => { + switch (params.eventId) { + case EventNavDropdownOption.RECOMMENDED: + case EventNavDropdownOption.LATEST: + case EventNavDropdownOption.OLDEST: + return params.eventId; + case undefined: + return EventNavDropdownOption.RECOMMENDED; + default: + return undefined; + } + }; + + const selectedValue = getSelectedOption(); + + return ( + <CompactSelect + size="sm" + options={EVENT_NAV_DROPDOWN_OPTIONS} + value={selectedValue} + triggerLabel={!selectedValue ? 'Navigate Events' : undefined} + onChange={selectedOption => { + switch (selectedOption.value) { + case EventNavDropdownOption.RECOMMENDED: + case EventNavDropdownOption.LATEST: + case EventNavDropdownOption.OLDEST: + browserHistory.push({ + pathname: normalizeUrl( + makeBaseEventsPath({organization, group}) + selectedOption.value + '/' + ), + query: {...location.query, referrer: `${selectedOption.value}-event`}, + }); + break; + case EventNavDropdownOption.ALL: + browserHistory.push({ + pathname: normalizeUrl( + `/organizations/${organization.slug}/issues/${group.id}/events/` + ), + query: location.query, + }); + break; + default: + break; + } + }} + /> + ); +} + export function GroupEventCarousel({event, group, projectSlug}: GroupEventCarouselProps) { const theme = useTheme(); const organization = useOrganization(); const location = useLocation(); - const largeViewport = useMedia(`(min-width: ${theme.breakpoints.large})`); const xlargeViewport = useMedia(`(min-width: ${theme.breakpoints.xlarge})`); const hasReplay = Boolean(event?.tags?.find(({key}) => key === 'replayId')?.value); @@ -147,50 +227,60 @@ export function GroupEventCarousel({event, group, projectSlug}: GroupEventCarous }); }; + const isHelpfulEventUiEnabled = + organization.features.includes('issue-details-most-helpful-event') && + organization.features.includes('issue-details-most-helpful-event-ui'); + return ( <CarouselAndButtonsWrapper> - <EventHeading> - <EventIdLabel>Event ID:</EventIdLabel>{' '} - <Button - aria-label={t('Copy')} - borderless - onClick={onClickCopy} - size="zero" - title={event.id} - tooltipProps={{overlayStyle: {maxWidth: 'max-content'}}} - translucentBorder - > - <EventId> - {getShortEventId(event.id)} - <CopyIconContainer> - <IconCopy size="xs" /> - </CopyIconContainer> - </EventId> - </Button>{' '} - {(event.dateCreated ?? event.dateReceived) && ( - <EventTimeLabel> - {getDynamicText({ - fixed: 'Jan 1, 12:00 AM', - value: ( - <Tooltip - isHoverable - showUnderline - title={<EventCreatedTooltip event={event} />} - overlayStyle={{maxWidth: 300}} - > - <DateTime date={event.dateCreated ?? event.dateReceived} /> - </Tooltip> - ), - })} - {isOverLatencyThreshold && ( - <Tooltip title="High latency"> - <StyledIconWarning size="xs" color="warningText" /> - </Tooltip> + <div> + <EventHeading> + <EventIdAndTimeContainer> + <EventIdContainer> + <strong>Event ID:</strong> + <Button + aria-label={t('Copy')} + borderless + onClick={onClickCopy} + size="zero" + title={event.id} + tooltipProps={{overlayStyle: {maxWidth: 'max-content'}}} + translucentBorder + > + <EventId> + {getShortEventId(event.id)} + <CopyIconContainer> + <IconCopy size="xs" /> + </CopyIconContainer> + </EventId> + </Button> + </EventIdContainer> + {(event.dateCreated ?? event.dateReceived) && ( + <EventTimeLabel> + {getDynamicText({ + fixed: 'Jan 1, 12:00 AM', + value: ( + <Tooltip + isHoverable + showUnderline + title={<EventCreatedTooltip event={event} />} + overlayStyle={{maxWidth: 300}} + > + <DateTime date={event.dateCreated ?? event.dateReceived} /> + </Tooltip> + ), + })} + {isOverLatencyThreshold && ( + <Tooltip title="High latency"> + <StyledIconWarning size="xs" color="warningText" /> + </Tooltip> + )} + </EventTimeLabel> )} - </EventTimeLabel> - )} + </EventIdAndTimeContainer> + </EventHeading> <QuickTrace event={event} organization={organization} location={location} /> - </EventHeading> + </div> <ActionsWrapper> <DropdownMenu position="bottom-end" @@ -216,7 +306,7 @@ export function GroupEventCarousel({event, group, projectSlug}: GroupEventCarous key: 'json', label: `JSON (${formatBytesBase2(event.size)})`, onAction: downloadJson, - hidden: largeViewport, + hidden: xlargeViewport, }, { key: 'full-event-discover', @@ -257,7 +347,7 @@ export function GroupEventCarousel({event, group, projectSlug}: GroupEventCarous Copy Link </Button> )} - {largeViewport && ( + {xlargeViewport && ( <Button size={BUTTON_SIZE} icon={<IconOpen size={BUTTON_ICON_SIZE} />} @@ -266,15 +356,18 @@ export function GroupEventCarousel({event, group, projectSlug}: GroupEventCarous JSON </Button> )} + <EventNavigationDropdown group={group} /> <NavButtons> - <EventNavigationButton - group={group} - icon={<IconPrevious size={BUTTON_ICON_SIZE} />} - disabled={!hasPreviousEvent} - title={t('First Event')} - eventId="oldest" - referrer="oldest-event" - /> + {!isHelpfulEventUiEnabled && ( + <EventNavigationButton + group={group} + icon={<IconPrevious size={BUTTON_ICON_SIZE} />} + disabled={!hasPreviousEvent} + title={t('First Event')} + eventId="oldest" + referrer="oldest-event" + /> + )} <EventNavigationButton group={group} icon={<IconChevron direction="left" size={BUTTON_ICON_SIZE} />} @@ -291,14 +384,16 @@ export function GroupEventCarousel({event, group, projectSlug}: GroupEventCarous eventId={event.nextEventID} referrer="next-event" /> - <EventNavigationButton - group={group} - icon={<IconNext size={BUTTON_ICON_SIZE} />} - disabled={!hasNextEvent} - title={t('Latest Event')} - eventId="latest" - referrer="latest-event" - /> + {!isHelpfulEventUiEnabled && ( + <EventNavigationButton + group={group} + icon={<IconNext size={BUTTON_ICON_SIZE} />} + disabled={!hasNextEvent} + title={t('Latest Event')} + eventId="latest" + referrer="latest-event" + /> + )} </NavButtons> </ActionsWrapper> </CarouselAndButtonsWrapper> @@ -314,6 +409,10 @@ const CarouselAndButtonsWrapper = styled('div')` `; const EventHeading = styled('div')` + display: flex; + align-items: center; + flex-wrap: wrap; + gap: ${space(1)}; font-size: ${p => p.theme.fontSizeLarge}; @media (max-width: 600px) { @@ -355,8 +454,18 @@ const NavButtons = styled('div')` } `; -const EventIdLabel = styled('span')` - font-weight: bold; +const EventIdAndTimeContainer = styled('div')` + display: flex; + align-items: center; + column-gap: ${space(0.75)}; + row-gap: 0; + flex-wrap: wrap; +`; + +const EventIdContainer = styled('div')` + display: flex; + align-items: center; + column-gap: ${space(0.25)}; `; const EventTimeLabel = styled('span')` @@ -378,6 +487,9 @@ const EventId = styled('span')` display: flex; } } + @media (max-width: 600px) { + font-size: ${p => p.theme.fontSizeMedium}; + } `; const CopyIconContainer = styled('span')`
fd99887982f76d67d6957f6acc06b8da6f8a91de
2024-12-21 01:25:36
Burak Yigit Kaya
fix(web): Add react_config context on auth pages take 2 (#82480)
false
Add react_config context on auth pages take 2 (#82480)
fix
diff --git a/src/sentry/web/frontend/auth_login.py b/src/sentry/web/frontend/auth_login.py index 302f1e7d69ff3a..cf19698091d73c 100644 --- a/src/sentry/web/frontend/auth_login.py +++ b/src/sentry/web/frontend/auth_login.py @@ -432,7 +432,7 @@ def get_ratelimited_login_form( ] metrics.incr("login.attempt", instance="rate_limited", skip_internal=True, sample_rate=1.0) - context = { + context = self.get_default_context(request=request) | { "op": "login", "login_form": login_form, "referrer": request.GET.get("referrer"), @@ -527,9 +527,7 @@ def get_default_context(self, request: Request, **kwargs) -> dict: default_context = { "server_hostname": get_server_hostname(), "login_form": None, - "organization": kwargs.pop( - "organization", None - ), # NOTE: not utilized in basic login page (only org login) + "organization": organization, # NOTE: not utilized in basic login page (only org login) "register_form": None, "CAN_REGISTER": False, "react_config": get_client_config(request, self.active_organization), @@ -704,18 +702,11 @@ def handle_basic_auth(self, request: Request, **kwargs) -> HttpResponseBase: "login.attempt", instance="failure", skip_internal=True, sample_rate=1.0 ) - context = { + context = self.get_default_context(request=request, organization=organization) | { "op": op or "login", - "server_hostname": get_server_hostname(), "login_form": login_form, - "organization": organization, "register_form": register_form, "CAN_REGISTER": can_register, - "join_request_link": self.get_join_request_link( - organization=organization, request=request - ), - "show_login_banner": settings.SHOW_LOGIN_BANNER, - "referrer": request.GET.get("referrer"), } context.update(additional_context.run_callbacks(request)) diff --git a/src/sentry/web/frontend/auth_organization_login.py b/src/sentry/web/frontend/auth_organization_login.py index 3e2de690770ea6..18af32980bcdc5 100644 --- a/src/sentry/web/frontend/auth_organization_login.py +++ b/src/sentry/web/frontend/auth_organization_login.py @@ -23,14 +23,15 @@ def respond_login(self, request: Request, context, *args, **kwargs) -> HttpRespo return self.respond("sentry/organization-login.html", context) def handle_sso(self, request: Request, organization: RpcOrganization, auth_provider): - referrer = request.GET.get("referrer") if request.method == "POST": helper = AuthHelper( request=request, organization=organization, auth_provider=auth_provider, flow=AuthHelper.FLOW_LOGIN, - referrer=referrer, # TODO: get referrer from the form submit - not the query parms + referrer=request.GET.get( + "referrer" + ), # TODO: get referrer from the form submit - not the query parms ) if request.POST.get("init"): @@ -47,13 +48,10 @@ def handle_sso(self, request: Request, organization: RpcOrganization, auth_provi provider = auth_provider.get_provider() - context = { - "CAN_REGISTER": False, - "organization": organization, + context = self.get_default_context(request, organization=organization) | { "provider_key": provider.key, "provider_name": provider.name, "authenticated": request.user.is_authenticated, - "referrer": referrer, } return self.respond("sentry/organization-login.html", context) diff --git a/src/sentry/web/frontend/oauth_authorize.py b/src/sentry/web/frontend/oauth_authorize.py index e9ed3e3543fc99..163b2f973440c6 100644 --- a/src/sentry/web/frontend/oauth_authorize.py +++ b/src/sentry/web/frontend/oauth_authorize.py @@ -233,13 +233,14 @@ def get(self, request: HttpRequest, **kwargs) -> HttpResponseBase: # If application is not org level we should not show organizations to choose from at all organization_options = [] - context = { + context = self.get_default_context(request) | { "user": request.user, "application": application, "scopes": scopes, "permissions": permissions, "organization_options": organization_options, } + return self.respond("sentry/oauth-authorize.html", context) def post(self, request: HttpRequest, **kwargs) -> HttpResponseBase: diff --git a/tests/sentry/web/frontend/test_react_page.py b/tests/sentry/web/frontend/test_react_page.py index d9f25f12b561e9..bd294f459075e1 100644 --- a/tests/sentry/web/frontend/test_react_page.py +++ b/tests/sentry/web/frontend/test_react_page.py @@ -296,14 +296,11 @@ def _run_customer_domain_elevated_privileges(self, is_superuser: bool, is_staff: assert response.redirect_chain == [ (f"http://{other_org.slug}.testserver/issues/", 302) ] + assert self.client.session["activeorg"] == other_org.slug else: assert response.redirect_chain == [ (f"http://{other_org.slug}.testserver/auth/login/{other_org.slug}/", 302) ] - - if is_superuser or is_staff: - assert self.client.session["activeorg"] == other_org.slug - else: assert "activeorg" not in self.client.session # Accessing org without customer domain as superuser and/or staff. diff --git a/tests/sentry/web/test_api.py b/tests/sentry/web/test_api.py index f1285d02730fdb..3b14b44003b357 100644 --- a/tests/sentry/web/test_api.py +++ b/tests/sentry/web/test_api.py @@ -8,7 +8,6 @@ from sentry import options from sentry.api.utils import generate_region_url from sentry.auth import superuser -from sentry.conf.types.sentry_config import SentryMode from sentry.deletions.models.scheduleddeletion import RegionScheduledDeletion from sentry.deletions.tasks.scheduled import run_deletion from sentry.models.apitoken import ApiToken @@ -78,61 +77,10 @@ class RobotsTxtTest(TestCase): def path(self): return reverse("sentry-robots-txt") - def test_robots_self_hosted(self): - with override_settings(SENTRY_MODE=SentryMode.SELF_HOSTED): - resp = self.client.get(self.path) - assert resp.status_code == 200 - assert resp["Content-Type"] == "text/plain" - assert ( - resp.content - == b"""\ -User-agent: * -Disallow: / -""" - ) - - def test_robots_saas(self): - with override_settings(SENTRY_MODE=SentryMode.SAAS): - resp = self.client.get(self.path) - assert resp.status_code == 200 - assert resp["Content-Type"] == "text/plain" - # This is sentry.io/robots.txt. - assert ( - resp.content - == b"""\ -User-agent: * -Disallow: /api/ -Allow: /api/*/store/ -Allow: / - -Sitemap: https://sentry.io/sitemap-index.xml -""" - ) - - # SaaS customer domains should disallow all. - resp = self.client.get(self.path, HTTP_HOST="foo.testserver") - assert resp.status_code == 200 - assert resp["Content-Type"] == "text/plain" - assert ( - resp.content - == b"""\ -User-agent: * -Disallow: / -""" - ) - - def test_robots_single_tenant(self): - with override_settings(SENTRY_MODE=SentryMode.SINGLE_TENANT): - resp = self.client.get(self.path) - assert resp.status_code == 200 - assert resp["Content-Type"] == "text/plain" - assert ( - resp.content - == b"""\ -User-agent: * -Disallow: / -""" - ) + def test_robots(self): + resp = self.client.get(self.path) + assert resp.status_code == 200 + assert resp["Content-Type"] == "text/plain" @region_silo_test(regions=create_test_regions("us", "eu"), include_monolith_run=True) @@ -301,13 +249,11 @@ def _run_test_with_privileges(self, is_superuser: bool, is_staff: bool): assert response.redirect_chain == [ (f"http://{other_org.slug}.testserver/issues/", 302) ] + assert self.client.session["activeorg"] == other_org.slug else: assert response.redirect_chain == [ (f"http://{other_org.slug}.testserver/auth/login/{other_org.slug}/", 302) ] - if is_superuser or is_staff: - assert self.client.session["activeorg"] == other_org.slug - else: assert "activeorg" not in self.client.session # lastOrganization is set
b57faf5061fc5b1ddcdd55def61abcb026eedc0d
2023-07-20 00:56:24
anthony sottile
ref: fix some mypy issues (#53158)
false
fix some mypy issues (#53158)
ref
diff --git a/pyproject.toml b/pyproject.toml index 0ef0ea3dc9daa7..f33e2831af588c 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -266,7 +266,6 @@ module = [ "sentry.api.endpoints.organization_searches", "sentry.api.endpoints.organization_sentry_function_details", "sentry.api.endpoints.organization_sessions", - "sentry.api.endpoints.organization_slugs", "sentry.api.endpoints.organization_stats", "sentry.api.endpoints.organization_tagkey_values", "sentry.api.endpoints.organization_tags", @@ -327,7 +326,6 @@ module = [ "sentry.api.endpoints.user_subscriptions", "sentry.api.event_search", "sentry.api.fields.actor", - "sentry.api.fields.user", "sentry.api.helpers.deprecation", "sentry.api.helpers.events", "sentry.api.helpers.group_index.delete", @@ -601,12 +599,6 @@ module = [ "sentry.issues.occurrence_consumer", "sentry.issues.search", "sentry.issues.status_change", - "sentry.lang.java.plugin", - "sentry.lang.javascript.errorlocale", - "sentry.lang.javascript.errormapping", - "sentry.lang.native.error", - "sentry.lang.native.processing", - "sentry.lang.native.symbolicator", "sentry.lint.engine", "sentry.mail.adapter", "sentry.mail.forms.assigned_to", @@ -623,7 +615,6 @@ module = [ "sentry.middleware.integrations.integration_control", "sentry.middleware.integrations.parsers.base", "sentry.middleware.integrations.parsers.slack", - "sentry.middleware.locale", "sentry.middleware.ratelimit", "sentry.middleware.subdomain", "sentry.middleware.superuser", @@ -644,22 +635,7 @@ module = [ "sentry.models.artifactbundle", "sentry.models.auditlogentry", "sentry.models.authprovider", - "sentry.models.avatars.base", - "sentry.models.avatars.doc_integration_avatar", - "sentry.models.avatars.organization_avatar", - "sentry.models.avatars.project_avatar", - "sentry.models.avatars.sentry_app_avatar", - "sentry.models.avatars.team_avatar", - "sentry.models.avatars.user_avatar", "sentry.models.commit", - "sentry.models.dashboard", - "sentry.models.dashboard_widget", - "sentry.models.files.abstractfile", - "sentry.models.files.abstractfileblob", - "sentry.models.files.control_file", - "sentry.models.files.control_fileblob", - "sentry.models.files.file", - "sentry.models.files.fileblob", "sentry.models.group", "sentry.models.groupassignee", "sentry.models.grouphistory", @@ -1329,7 +1305,6 @@ module = [ "tests.sentry.services.test_http", "tests.sentry.shared_integrations.client.test_base", "tests.sentry.snuba.metrics.fields.test_base", - "tests.sentry.snuba.metrics.test_metrics_layer.test_metrics_enhanced_performance", "tests.sentry.snuba.metrics.test_metrics_layer.test_release_health", "tests.sentry.snuba.metrics.test_mqb_query_transformer", "tests.sentry.snuba.metrics.test_query_builder", @@ -1419,7 +1394,6 @@ module = [ "tests.snuba.api.serializers.test_group_stream", "tests.snuba.incidents.test_tasks", "tests.snuba.rules.conditions.test_event_frequency", - "tests.snuba.search.test_backend", "tests.snuba.sessions.test_sessions", "tests.snuba.sessions.test_sessions_v2", "tests.snuba.tagstore.test_tagstore_backend", diff --git a/src/sentry/api/endpoints/organization_slugs.py b/src/sentry/api/endpoints/organization_slugs.py index 151732dd638f30..e7e978d266b5a1 100644 --- a/src/sentry/api/endpoints/organization_slugs.py +++ b/src/sentry/api/endpoints/organization_slugs.py @@ -1,4 +1,5 @@ -from django.core.validators import ValidationError, validate_slug +from django.core.exceptions import ValidationError +from django.core.validators import validate_slug from django.db import router, transaction from rest_framework.request import Request from rest_framework.response import Response diff --git a/src/sentry/api/event_search.py b/src/sentry/api/event_search.py index a12fa65165817b..5100740f403276 100644 --- a/src/sentry/api/event_search.py +++ b/src/sentry/api/event_search.py @@ -1,3 +1,5 @@ +from __future__ import annotations + import re from collections import namedtuple from dataclasses import asdict, dataclass, field @@ -466,7 +468,7 @@ class SearchConfig: free_text_key = "message" @classmethod - def create_from(cls, search_config: "SearchConfig", **overrides): + def create_from(cls, search_config: SearchConfig, **overrides): config = cls(**asdict(search_config)) for key, val in overrides.items(): setattr(config, key, val) @@ -1136,7 +1138,7 @@ def generic_visit(self, node, children): def parse_search_query( query, config=None, params=None, builder=None, config_overrides=None -) -> Sequence[SearchFilter]: +) -> list[SearchFilter]: if config is None: config = default_config diff --git a/src/sentry/api/fields/user.py b/src/sentry/api/fields/user.py index f39bdc893d94b6..1b30d5ef3c826f 100644 --- a/src/sentry/api/fields/user.py +++ b/src/sentry/api/fields/user.py @@ -4,21 +4,22 @@ from rest_framework import serializers +from sentry.models.user import User from sentry.services.hybrid_cloud.user import RpcUser from sentry.services.hybrid_cloud.user.service import user_service from sentry.utils.auth import find_users class UserField(serializers.Field): - def to_representation(self, value: RpcUser): + def to_representation(self, value: RpcUser) -> str: return value.username - def to_internal_value(self, data: Any) -> RpcUser | None: + def to_internal_value(self, data: Any) -> RpcUser | User | None: if not data: return None if isinstance(data, int) or data.isdigit(): - user: RpcUser = user_service.get_user(user_id=data) + user = user_service.get_user(user_id=data) if user is not None: return user diff --git a/src/sentry/db/router.py b/src/sentry/db/router.py index 807afba2d126af..6545033b1298e9 100644 --- a/src/sentry/db/router.py +++ b/src/sentry/db/router.py @@ -6,7 +6,7 @@ from django.apps import apps from django.db import connections -from django.db.utils import ConnectionDoesNotExist +from django.utils.connection import ConnectionDoesNotExist from sentry.db.models.base import Model, ModelSiloLimit from sentry.silo.base import SiloLimit, SiloMode diff --git a/src/sentry/lang/javascript/errorlocale.py b/src/sentry/lang/javascript/errorlocale.py index c0c4bde09adc32..d7188ea9ef8b09 100644 --- a/src/sentry/lang/javascript/errorlocale.py +++ b/src/sentry/lang/javascript/errorlocale.py @@ -7,10 +7,10 @@ TARGET_LOCALE = "en-US" translation_lookup_table = set() -target_locale_lookup_table = dict() +target_locale_lookup_table = {} -def populate_target_locale_lookup_table(): +def populate_target_locale_lookup_table() -> None: for locale in os.listdir(LOCALES_DIR): fn = os.path.join(LOCALES_DIR, locale) if not os.path.isfile(fn): @@ -29,8 +29,8 @@ def populate_target_locale_lookup_table(): r"%s", r"(?P<format_string_data>[a-zA-Z0-9-_\$]+)" ) # Some errors are substrings of more detailed ones, so we need exact match - translation_regexp = re.compile("^" + translation_regexp + "$") - translation_lookup_table.add((translation_regexp, key)) + translation_regexp_re = re.compile(f"^{translation_regexp}$") + translation_lookup_table.add((translation_regexp_re, key)) def find_translation(message): @@ -71,8 +71,8 @@ def translate_message(original_message): match = message_type_regexp.search(message) if match is not None: - type = match.groupdict().get("type") - message = match.groupdict().get("message") + type = match["type"] + message = match["message"] translation, format_string_data = find_translation(message) diff --git a/src/sentry/lang/javascript/errormapping.py b/src/sentry/lang/javascript/errormapping.py index 5b5bfd662b3b9d..80fbf72b2c87f1 100644 --- a/src/sentry/lang/javascript/errormapping.py +++ b/src/sentry/lang/javascript/errormapping.py @@ -1,3 +1,5 @@ +from __future__ import annotations + import logging import random import re @@ -23,7 +25,7 @@ "https://raw.githubusercontent.com/facebook/" "react/master/scripts/error-codes/codes.json" ) -error_processors = {} +error_processors: dict[str, Processor] = {} def is_expired(ts): @@ -95,15 +97,13 @@ def process_react_exception(exc, match, mapping): args = [] for k, v in parse_qsl(qs, keep_blank_values=True): if k == "args[]": - if isinstance(v, bytes): - v = v.decode("utf-8", "replace") args.append(v) # Due to truncated error messages we sometimes might not be able to # get all arguments. In that case we fill up missing parameters for # the format string with <redacted>. - args = tuple(args + ["<redacted>"] * (arg_count - len(args)))[:arg_count] - exc["value"] = msg_format % args + args_t = tuple(args + ["<redacted>"] * (arg_count - len(args)))[:arg_count] + exc["value"] = msg_format % args_t return True @@ -121,7 +121,6 @@ def rewrite_exception(data): if exc is None: continue - processor: Processor for processor in error_processors.values(): try: original_value = exc.get("value") diff --git a/src/sentry/lang/native/error.py b/src/sentry/lang/native/error.py index 668ad6cabce7bb..0f4af17430042c 100644 --- a/src/sentry/lang/native/error.py +++ b/src/sentry/lang/native/error.py @@ -1,3 +1,5 @@ +from __future__ import annotations + import logging from sentry.lang.native.utils import image_name, is_minidump_event @@ -33,14 +35,14 @@ def __init__(self, message=None, type=None, obj=None): Exception.__init__(self) self.message = str(message) self.type = type - self.image_name = None - self.image_path = None + self.image_name: str | None = None + self.image_path: str | None = None if obj is not None: - self.image_uuid = str(obj.debug_id) + self.image_uuid: str | None = str(obj.debug_id) if obj.name: self.image_path = obj.name self.image_name = image_name(obj.name) - self.image_arch = obj.arch + self.image_arch: str | None = obj.arch else: self.image_uuid = None self.image_arch = None diff --git a/src/sentry/lang/native/processing.py b/src/sentry/lang/native/processing.py index 9d09831e848401..bc2e6e81541c45 100644 --- a/src/sentry/lang/native/processing.py +++ b/src/sentry/lang/native/processing.py @@ -1,3 +1,5 @@ +from __future__ import annotations + import logging import posixpath from typing import Any, Callable, Optional, Set @@ -205,11 +207,11 @@ def _merge_full_response(data, response): os = get_os_from_event(data) - images = [] + images: list[dict[str, Any]] = [] set_path(data, "debug_meta", "images", value=images) for complete_image in response["modules"]: - image = {} + image: dict[str, Any] = {} _merge_image(image, complete_image, os, data) images.append(image) @@ -230,7 +232,7 @@ def _merge_full_response(data, response): if response.get("crash_reason"): data_exception["type"] = response["crash_reason"] - data_threads = [] + data_threads: list[dict[str, Any]] = [] if response["stacktraces"]: data["threads"] = {"values": data_threads} else: @@ -266,7 +268,7 @@ def _merge_full_response(data, response): data_stacktrace["registers"] = complete_stacktrace["registers"] for complete_frame in reversed(complete_stacktrace["frames"]): - new_frame = {} + new_frame: dict[str, Any] = {} _merge_frame(new_frame, complete_frame) data_stacktrace["frames"].append(new_frame) @@ -412,7 +414,7 @@ def process_native_stacktraces(symbolicator: Symbolicator, data: Any) -> Any: assert len(stacktraces) == len(response["stacktraces"]), (stacktraces, response) for sinfo, complete_stacktrace in zip(stacktrace_infos, response["stacktraces"]): - complete_frames_by_idx = {} + complete_frames_by_idx: dict[int, list[dict[str, Any]]] = {} for complete_frame in complete_stacktrace.get("frames") or (): complete_frames_by_idx.setdefault(complete_frame["original_index"], []).append( complete_frame diff --git a/src/sentry/lang/native/symbolicator.py b/src/sentry/lang/native/symbolicator.py index c5a3a7237bd770..118fda0f319211 100644 --- a/src/sentry/lang/native/symbolicator.py +++ b/src/sentry/lang/native/symbolicator.py @@ -236,13 +236,10 @@ def close(self): self.session.close() self.session = None - def _ensure_open(self): + def _request(self, method, path, **kwargs): if not self.session: raise RuntimeError("Session not opened") - def _request(self, method, path, **kwargs): - self._ensure_open() - url = urljoin(self.url, path) # required for load balancing diff --git a/src/sentry/middleware/locale.py b/src/sentry/middleware/locale.py index 351485dd862f6c..12d55a3418f73c 100644 --- a/src/sentry/middleware/locale.py +++ b/src/sentry/middleware/locale.py @@ -1,17 +1,17 @@ import pytz import sentry_sdk from django.conf import settings +from django.http.request import HttpRequest +from django.http.response import HttpResponseBase from django.middleware.locale import LocaleMiddleware -from django.utils.translation import LANGUAGE_SESSION_KEY, _trans -from rest_framework.request import Request -from rest_framework.response import Response +from django.utils import translation from sentry.services.hybrid_cloud.user_option import get_option_from_list, user_option_service from sentry.utils.safe import safe_execute class SentryLocaleMiddleware(LocaleMiddleware): - def process_request(self, request: Request): + def process_request(self, request: HttpRequest) -> None: with sentry_sdk.start_span(op="middleware.locale", description="process_request"): # No locale for static media # This avoids touching user session, which means we avoid @@ -27,16 +27,16 @@ def process_request(self, request: Request): # user is explicitly forcing language if lang_code: try: - language = _trans.get_supported_language_variant(lang_code) + language = translation.get_supported_language_variant(lang_code) except LookupError: super().process_request(request) else: - _trans.activate(language) - request.LANGUAGE_CODE = _trans.get_language() + translation.activate(language) + request.LANGUAGE_CODE = translation.get_language() else: super().process_request(request) - def load_user_conf(self, request: Request): + def load_user_conf(self, request: HttpRequest) -> None: if not request.user.is_authenticated: return @@ -45,12 +45,14 @@ def load_user_conf(self, request: Request): ) if language := get_option_from_list(options, key="language"): - request.session[LANGUAGE_SESSION_KEY] = language + request.session[translation.LANGUAGE_SESSION_KEY] = language if timezone := get_option_from_list(options, key="timezone"): - request.timezone = pytz.timezone(timezone) + request.timezone = pytz.timezone(timezone) # type: ignore[attr-defined] - def process_response(self, request: Request, response: Response) -> Response: + def process_response( + self, request: HttpRequest, response: HttpResponseBase + ) -> HttpResponseBase: # If static bound, we don't want to run the normal process_response since this # adds an extra `Vary: Accept-Language`. Static files don't need this and is # less effective for caching. diff --git a/src/sentry/models/avatars/base.py b/src/sentry/models/avatars/base.py index 08c0340c9798d4..79ce6e9868cb19 100644 --- a/src/sentry/models/avatars/base.py +++ b/src/sentry/models/avatars/base.py @@ -1,4 +1,7 @@ +from __future__ import annotations + from io import BytesIO +from typing import ClassVar from uuid import uuid4 from django.core.exceptions import ObjectDoesNotExist @@ -22,9 +25,11 @@ class AvatarBase(Model): __include_in_export__ = False - ALLOWED_SIZES = (20, 32, 36, 48, 52, 64, 80, 96, 120) + ALLOWED_SIZES: ClassVar[tuple[int, ...]] = (20, 32, 36, 48, 52, 64, 80, 96, 120) - FILE_TYPE = None + # abstract + AVATAR_TYPES: ClassVar[tuple[tuple[int, str], ...]] + FILE_TYPE: ClassVar[str] file_id = BoundedBigIntegerField(unique=True, null=True) ident = models.CharField(max_length=32, unique=True, db_index=True) diff --git a/src/sentry/models/avatars/doc_integration_avatar.py b/src/sentry/models/avatars/doc_integration_avatar.py index 3efdef97d742a5..c669600ea69cf0 100644 --- a/src/sentry/models/avatars/doc_integration_avatar.py +++ b/src/sentry/models/avatars/doc_integration_avatar.py @@ -1,7 +1,5 @@ from __future__ import annotations -from typing import ClassVar - from django.db import models from django.utils import timezone @@ -16,9 +14,9 @@ class DocIntegrationAvatar(ControlAvatarBase): A DocIntegrationAvatar associates a DocIntegration with a logo photo File. """ - AVATAR_TYPES: ClassVar[tuple[tuple[int, str], ...]] = ((0, "upload"),) + AVATAR_TYPES = ((0, "upload"),) - FILE_TYPE: ClassVar[str] = "avatar.file" + FILE_TYPE = "avatar.file" doc_integration = FlexibleForeignKey("sentry.DocIntegration", related_name="avatar") avatar_type = models.PositiveSmallIntegerField(default=0, choices=((0, "upload"),)) diff --git a/src/sentry/models/avatars/organization_avatar.py b/src/sentry/models/avatars/organization_avatar.py index 7523eea131354b..a331926f1d5414 100644 --- a/src/sentry/models/avatars/organization_avatar.py +++ b/src/sentry/models/avatars/organization_avatar.py @@ -1,7 +1,5 @@ from __future__ import annotations -from typing import ClassVar - from django.db import models from sentry.db.models import FlexibleForeignKey, region_silo_only_model @@ -16,9 +14,9 @@ class OrganizationAvatar(AvatarBase): and contains their preferences for avatar type. """ - AVATAR_TYPES: ClassVar[tuple[tuple[int, str], ...]] = ((0, "letter_avatar"), (1, "upload")) + AVATAR_TYPES = ((0, "letter_avatar"), (1, "upload")) - FILE_TYPE: ClassVar[str] = "avatar.file" + FILE_TYPE = "avatar.file" organization = FlexibleForeignKey("sentry.Organization", unique=True, related_name="avatar") avatar_type = models.PositiveSmallIntegerField(default=0, choices=AVATAR_TYPES) diff --git a/src/sentry/models/avatars/project_avatar.py b/src/sentry/models/avatars/project_avatar.py index 2b5217b78113a4..f5e347d45316cd 100644 --- a/src/sentry/models/avatars/project_avatar.py +++ b/src/sentry/models/avatars/project_avatar.py @@ -1,7 +1,5 @@ from __future__ import annotations -from typing import ClassVar - from django.db import models from sentry.db.models import FlexibleForeignKey, region_silo_only_model @@ -16,9 +14,9 @@ class ProjectAvatar(AvatarBase): and contains their preferences for avatar type. """ - AVATAR_TYPES: ClassVar[tuple[tuple[int, str], ...]] = ((0, "letter_avatar"), (1, "upload")) + AVATAR_TYPES = ((0, "letter_avatar"), (1, "upload")) - FILE_TYPE: ClassVar[str] = "avatar.file" + FILE_TYPE = "avatar.file" project = FlexibleForeignKey("sentry.Project", unique=True, related_name="avatar") avatar_type = models.PositiveSmallIntegerField(default=0, choices=AVATAR_TYPES) diff --git a/src/sentry/models/avatars/sentry_app_avatar.py b/src/sentry/models/avatars/sentry_app_avatar.py index 3e34a7cbed2417..aed7c4d62271a5 100644 --- a/src/sentry/models/avatars/sentry_app_avatar.py +++ b/src/sentry/models/avatars/sentry_app_avatar.py @@ -2,7 +2,7 @@ from collections import defaultdict from enum import Enum -from typing import TYPE_CHECKING, ClassVar, List +from typing import TYPE_CHECKING, List from django.db import models @@ -47,7 +47,7 @@ class SentryAppAvatar(ControlAvatarBase): AVATAR_TYPES = SentryAppAvatarTypes.get_choices() - FILE_TYPE: ClassVar[str] = "avatar.file" + FILE_TYPE = "avatar.file" sentry_app = FlexibleForeignKey("sentry.SentryApp", related_name="avatar") avatar_type = models.PositiveSmallIntegerField(default=0, choices=AVATAR_TYPES) diff --git a/src/sentry/models/avatars/team_avatar.py b/src/sentry/models/avatars/team_avatar.py index 8d63951079a491..38b369f7acefb4 100644 --- a/src/sentry/models/avatars/team_avatar.py +++ b/src/sentry/models/avatars/team_avatar.py @@ -1,7 +1,5 @@ from __future__ import annotations -from typing import ClassVar - from django.db import models from sentry.db.models import FlexibleForeignKey, region_silo_only_model @@ -16,9 +14,9 @@ class TeamAvatar(AvatarBase): and contains their preferences for avatar type. """ - AVATAR_TYPES: ClassVar[tuple[tuple[int, str], ...]] = ((0, "letter_avatar"), (1, "upload")) + AVATAR_TYPES = ((0, "letter_avatar"), (1, "upload")) - FILE_TYPE: ClassVar[str] = "avatar.file" + FILE_TYPE = "avatar.file" team = FlexibleForeignKey("sentry.Team", unique=True, related_name="avatar") avatar_type = models.PositiveSmallIntegerField(default=0, choices=AVATAR_TYPES) diff --git a/src/sentry/models/avatars/user_avatar.py b/src/sentry/models/avatars/user_avatar.py index 7df787d9b9a922..5e7cec07dce756 100644 --- a/src/sentry/models/avatars/user_avatar.py +++ b/src/sentry/models/avatars/user_avatar.py @@ -1,7 +1,5 @@ from __future__ import annotations -from typing import ClassVar - from django.db import models from sentry.db.models import BaseManager, FlexibleForeignKey, control_silo_only_model @@ -16,13 +14,9 @@ class UserAvatar(ControlAvatarBase): and contains their preferences for avatar type. """ - AVATAR_TYPES: ClassVar[tuple[tuple[int, str], ...]] = ( - (0, "letter_avatar"), - (1, "upload"), - (2, "gravatar"), - ) + AVATAR_TYPES = ((0, "letter_avatar"), (1, "upload"), (2, "gravatar")) - FILE_TYPE: ClassVar[str] = "avatar.file" + FILE_TYPE = "avatar.file" user = FlexibleForeignKey("sentry.User", unique=True, related_name="avatar") avatar_type = models.PositiveSmallIntegerField(default=0, choices=AVATAR_TYPES) diff --git a/src/sentry/models/dashboard.py b/src/sentry/models/dashboard.py index 7e4b7039f9b055..5d59eda41c9b71 100644 --- a/src/sentry/models/dashboard.py +++ b/src/sentry/models/dashboard.py @@ -1,3 +1,7 @@ +from __future__ import annotations + +from typing import Any + from django.db import models from django.utils import timezone @@ -94,225 +98,223 @@ class Meta: # # All widgets and queries in prebuilt dashboards must not have id attributes defined, # or users will be unable to 'update' them with a forked version. -PREBUILT_DASHBOARDS = { - item["id"]: item - for item in [ - { - # This should match the general template in static/app/views/dashboardsV2/data.tsx - "id": "default-overview", - "title": "General", - "dateCreated": "", - "createdBy": "", - "widgets": [ - { - "title": "Number of Errors", - "displayType": "big_number", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "", - "conditions": "!event.type:transaction", - "fields": ["count()"], - "aggregates": ["count()"], - "columns": [], - "orderby": "", - } - ], - }, - { - "title": "Number of Issues", - "displayType": "big_number", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "", - "conditions": "!event.type:transaction", - "fields": ["count_unique(issue)"], - "aggregates": ["count_unique(issue)"], - "columns": [], - "orderby": "", - } - ], - }, - { - "title": "Events", - "displayType": "line", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "Events", - "conditions": "!event.type:transaction", - "fields": ["count()"], - "aggregates": ["count()"], - "columns": [], - "orderby": "", - } - ], - }, - { - "title": "Affected Users", - "displayType": "line", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "Known Users", - "conditions": "has:user.email !event.type:transaction", - "fields": ["count_unique(user)"], - "aggregates": ["count_unique(user)"], - "columns": [], - "orderby": "", - }, - { - "name": "Anonymous Users", - "conditions": "!has:user.email !event.type:transaction", - "fields": ["count_unique(user)"], - "aggregates": ["count_unique(user)"], - "columns": [], - "orderby": "", - }, - ], - }, - { - "title": "Handled vs. Unhandled", - "displayType": "line", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "Handled", - "conditions": "error.handled:true", - "fields": ["count()"], - "aggregates": ["count()"], - "columns": [], - "orderby": "", - }, - { - "name": "Unhandled", - "conditions": "error.handled:false", - "fields": ["count()"], - "aggregates": ["count()"], - "columns": [], - "orderby": "", - }, - ], - }, - { - "title": "Errors by Country", - "displayType": "table", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "Error counts", - "conditions": "!event.type:transaction has:geo.country_code", - "fields": ["geo.country_code", "geo.region", "count()"], - "aggregates": ["count()"], - "columns": ["geo.country_code", "geo.region"], - "orderby": "", - } - ], - }, - { - "title": "Errors by Browser", - "displayType": "table", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "", - "conditions": "!event.type:transaction has:browser.name", - "fields": ["browser.name", "count()"], - "aggregates": ["count()"], - "columns": ["browser.name"], - "orderby": "-count", - } - ], - }, - { - "title": "High Throughput Transactions", - "displayType": "table", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "", - "fields": ["count()", "transaction"], - "aggregates": ["count()"], - "columns": ["transaction"], - "conditions": "event.type:transaction", - "orderby": "-count()", - }, - ], - }, - { - "title": "Overall User Misery", - "displayType": "big_number", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "", - "fields": ["user_misery(300)"], - "aggregates": ["user_misery(300)"], - "columns": [], - "conditions": "", - "orderby": "", - }, - ], - }, - { - "title": "High Throughput Transactions", - "displayType": "top_n", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "", - "fields": ["transaction", "count()"], - "aggregates": ["count()"], - "columns": ["transaction"], - "conditions": "event.type:transaction", - "orderby": "-count()", - }, - ], - }, - { - "title": "Issues Assigned to Me or My Teams", - "displayType": "table", - "interval": "5m", - "queries": [ - { - "name": "", - "fields": ["assignee", "issue", "title"], - "aggregates": [], - "columns": ["assignee", "issue", "title"], - "conditions": "assigned_or_suggested:me is:unresolved", - "orderby": "priority", - }, - ], - "widgetType": "issue", - }, - { - "title": "Transactions Ordered by Misery", - "displayType": "table", - "interval": "5m", - "widgetType": "discover", - "queries": [ - { - "name": "", - "fields": ["transaction", "user_misery(300)"], - "aggregates": ["user_misery(300)"], - "columns": ["transaction"], - "conditions": "", - "orderby": "-user_misery(300)", - }, - ], - }, - ], - } - ] -} +_PREBUILT_DASHBOARDS: list[dict[str, Any]] = [ + { + # This should match the general template in static/app/views/dashboardsV2/data.tsx + "id": "default-overview", + "title": "General", + "dateCreated": "", + "createdBy": "", + "widgets": [ + { + "title": "Number of Errors", + "displayType": "big_number", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "", + "conditions": "!event.type:transaction", + "fields": ["count()"], + "aggregates": ["count()"], + "columns": [], + "orderby": "", + } + ], + }, + { + "title": "Number of Issues", + "displayType": "big_number", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "", + "conditions": "!event.type:transaction", + "fields": ["count_unique(issue)"], + "aggregates": ["count_unique(issue)"], + "columns": [], + "orderby": "", + } + ], + }, + { + "title": "Events", + "displayType": "line", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "Events", + "conditions": "!event.type:transaction", + "fields": ["count()"], + "aggregates": ["count()"], + "columns": [], + "orderby": "", + } + ], + }, + { + "title": "Affected Users", + "displayType": "line", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "Known Users", + "conditions": "has:user.email !event.type:transaction", + "fields": ["count_unique(user)"], + "aggregates": ["count_unique(user)"], + "columns": [], + "orderby": "", + }, + { + "name": "Anonymous Users", + "conditions": "!has:user.email !event.type:transaction", + "fields": ["count_unique(user)"], + "aggregates": ["count_unique(user)"], + "columns": [], + "orderby": "", + }, + ], + }, + { + "title": "Handled vs. Unhandled", + "displayType": "line", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "Handled", + "conditions": "error.handled:true", + "fields": ["count()"], + "aggregates": ["count()"], + "columns": [], + "orderby": "", + }, + { + "name": "Unhandled", + "conditions": "error.handled:false", + "fields": ["count()"], + "aggregates": ["count()"], + "columns": [], + "orderby": "", + }, + ], + }, + { + "title": "Errors by Country", + "displayType": "table", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "Error counts", + "conditions": "!event.type:transaction has:geo.country_code", + "fields": ["geo.country_code", "geo.region", "count()"], + "aggregates": ["count()"], + "columns": ["geo.country_code", "geo.region"], + "orderby": "", + } + ], + }, + { + "title": "Errors by Browser", + "displayType": "table", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "", + "conditions": "!event.type:transaction has:browser.name", + "fields": ["browser.name", "count()"], + "aggregates": ["count()"], + "columns": ["browser.name"], + "orderby": "-count", + } + ], + }, + { + "title": "High Throughput Transactions", + "displayType": "table", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "", + "fields": ["count()", "transaction"], + "aggregates": ["count()"], + "columns": ["transaction"], + "conditions": "event.type:transaction", + "orderby": "-count()", + }, + ], + }, + { + "title": "Overall User Misery", + "displayType": "big_number", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "", + "fields": ["user_misery(300)"], + "aggregates": ["user_misery(300)"], + "columns": [], + "conditions": "", + "orderby": "", + }, + ], + }, + { + "title": "High Throughput Transactions", + "displayType": "top_n", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "", + "fields": ["transaction", "count()"], + "aggregates": ["count()"], + "columns": ["transaction"], + "conditions": "event.type:transaction", + "orderby": "-count()", + }, + ], + }, + { + "title": "Issues Assigned to Me or My Teams", + "displayType": "table", + "interval": "5m", + "queries": [ + { + "name": "", + "fields": ["assignee", "issue", "title"], + "aggregates": [], + "columns": ["assignee", "issue", "title"], + "conditions": "assigned_or_suggested:me is:unresolved", + "orderby": "priority", + }, + ], + "widgetType": "issue", + }, + { + "title": "Transactions Ordered by Misery", + "displayType": "table", + "interval": "5m", + "widgetType": "discover", + "queries": [ + { + "name": "", + "fields": ["transaction", "user_misery(300)"], + "aggregates": ["user_misery(300)"], + "columns": ["transaction"], + "conditions": "", + "orderby": "-user_misery(300)", + }, + ], + }, + ], + } +] +PREBUILT_DASHBOARDS = {item["id"]: item for item in _PREBUILT_DASHBOARDS} diff --git a/src/sentry/models/dashboard_widget.py b/src/sentry/models/dashboard_widget.py index 3f2cfd6f9378cb..9c3b6b18a05bc6 100644 --- a/src/sentry/models/dashboard_widget.py +++ b/src/sentry/models/dashboard_widget.py @@ -18,7 +18,7 @@ class TypesClass: - TYPES = [] + TYPES: list[tuple[int, str]] @classmethod def as_choices(cls): diff --git a/src/sentry/models/debugfile.py b/src/sentry/models/debugfile.py index a77cd70159a6c0..f5e33a06ac7284 100644 --- a/src/sentry/models/debugfile.py +++ b/src/sentry/models/debugfile.py @@ -16,6 +16,7 @@ Any, BinaryIO, ClassVar, + Container, Dict, FrozenSet, Iterable, @@ -85,7 +86,7 @@ def find_missing(self, checksums: Iterable[str], project: Project) -> List[str]: return sorted(missing) def find_by_debug_ids( - self, project: Project, debug_ids: List[str], features: Iterable[str] | None = None + self, project: Project, debug_ids: Container[str], features: Iterable[str] | None = None ) -> Dict[str, ProjectDebugFile]: """Finds debug information files matching the given debug identifiers. diff --git a/src/sentry/models/files/abstractfile.py b/src/sentry/models/files/abstractfile.py index f1064e80e21686..3933f722c33718 100644 --- a/src/sentry/models/files/abstractfile.py +++ b/src/sentry/models/files/abstractfile.py @@ -1,16 +1,21 @@ +from __future__ import annotations + import io import mmap import os import tempfile from concurrent.futures import ThreadPoolExecutor from hashlib import sha1 +from typing import ClassVar, Type from django.core.files.base import ContentFile from django.core.files.base import File as FileObj from django.db import models, router, transaction from django.utils import timezone +from sentry.celery import SentryTask from sentry.db.models import BoundedPositiveIntegerField, JSONField, Model +from sentry.models.files.abstractfileblob import AbstractFileBlob from sentry.models.files.utils import DEFAULT_BLOB_SIZE, AssembleChecksumMismatch, nooplogger from sentry.utils import metrics from sentry.utils.db import atomic_transaction @@ -39,6 +44,7 @@ def __exit__(self, exc_type, exc_value, tb): def detach_tempfile(self): if not self.prefetched: raise TypeError("Can only detech tempfiles in prefetch mode") + assert self._curfile is not None rv = self._curfile self._curfile = None self.close() @@ -109,6 +115,7 @@ def _seek(self, pos): raise ValueError("I/O operation on closed file") if self.prefetched: + assert self._curfile is not None return self._curfile.seek(pos) if pos < 0: @@ -125,6 +132,8 @@ def _seek(self, pos): break else: raise ValueError("Cannot seek to pos") + assert self._curfile is not None + assert self._curidx is not None self._curfile.seek(pos - self._curidx.offset) def seek(self, pos, whence=io.SEEK_SET): @@ -141,9 +150,12 @@ def tell(self): if self.closed: raise ValueError("I/O operation on closed file") if self.prefetched: + assert self._curfile is not None return self._curfile.tell() if self._curfile is None: return self.size + assert self._curidx is not None + assert self._curfile is not None return self._curidx.offset + self._curfile.tell() def read(self, n=-1): @@ -151,6 +163,7 @@ def read(self, n=-1): raise ValueError("I/O operation on closed file") if self.prefetched: + assert self._curfile is not None return self._curfile.read(n) result = bytearray() @@ -190,9 +203,11 @@ class AbstractFile(Model): class Meta: abstract = True - FILE_BLOB_MODEL = None - FILE_BLOB_INDEX_MODEL = None - DELETE_UNREFERENCED_BLOB_TASK = None + # abstract + FILE_BLOB_MODEL: ClassVar[Type[AbstractFileBlob]] + FILE_BLOB_INDEX_MODEL: ClassVar[Type[Model]] + DELETE_UNREFERENCED_BLOB_TASK: ClassVar[SentryTask] + blobs: models.ManyToManyField def _get_chunked_blob(self, mode=None, prefetch=False, prefetch_to=None, delete=True): return ChunkedFileBlobIndexWrapper( diff --git a/src/sentry/models/files/abstractfileblob.py b/src/sentry/models/files/abstractfileblob.py index d99937c5e3433d..ff5331b8615162 100644 --- a/src/sentry/models/files/abstractfileblob.py +++ b/src/sentry/models/files/abstractfileblob.py @@ -1,12 +1,18 @@ +from __future__ import annotations + +from abc import abstractmethod from concurrent.futures import ThreadPoolExecutor from threading import Semaphore +from typing import Any, ClassVar from uuid import uuid4 from django.db import IntegrityError, models, router from django.utils import timezone +from sentry.celery import SentryTask from sentry.db.models import BoundedPositiveIntegerField, Model from sentry.locks import locks +from sentry.models.files.abstractfileblobowner import AbstractFileBlobOwner from sentry.models.files.utils import ( UPLOAD_RETRY_TIME, _get_size_and_checksum, @@ -32,8 +38,14 @@ class AbstractFileBlob(Model): class Meta: abstract = True - FILE_BLOB_OWNER_MODEL = None - DELETE_FILE_TASK = None + # abstract + FILE_BLOB_OWNER_MODEL: ClassVar[type[AbstractFileBlobOwner]] + DELETE_FILE_TASK: ClassVar[SentryTask] + + @classmethod + @abstractmethod + def _storage_config(cls) -> dict[str, Any] | None: + raise NotImplementedError(cls) @classmethod def from_files(cls, files, organization=None, logger=nooplogger): diff --git a/src/sentry/snuba/metrics/query.py b/src/sentry/snuba/metrics/query.py index 7d7a9332206c35..5d4a82ab84e455 100644 --- a/src/sentry/snuba/metrics/query.py +++ b/src/sentry/snuba/metrics/query.py @@ -1,4 +1,6 @@ """ Classes needed to build a metrics query. Inspired by snuba_sdk.query. """ +from __future__ import annotations + from collections.abc import Mapping from dataclasses import dataclass from datetime import datetime, timedelta @@ -35,7 +37,9 @@ class MetricField: op: Optional[MetricOperationType] metric_mri: str - params: Optional[Dict[str, Union[str, int, float, Sequence[Tuple[Union[str, int]]]]]] = None + params: Optional[ + Dict[str, Union[None, str, int, float, Sequence[Tuple[Union[str, int], ...]]]] + ] = None alias: Optional[str] = None def __post_init__(self) -> None: diff --git a/src/sentry/templatetags/sentry_helpers.py b/src/sentry/templatetags/sentry_helpers.py index 6a0e4f9b12caad..2bd181e5b9ee5f 100644 --- a/src/sentry/templatetags/sentry_helpers.py +++ b/src/sentry/templatetags/sentry_helpers.py @@ -217,7 +217,7 @@ def get_sentry_version(context): @register.filter def timesince(value, now=None): - from django.template.defaultfilters import timesince + from django.utils.timesince import timesince if now is None: now = timezone.now() diff --git a/tests/sentry/snuba/metrics/test_metrics_layer/test_metrics_enhanced_performance.py b/tests/sentry/snuba/metrics/test_metrics_layer/test_metrics_enhanced_performance.py index f6c3439afd03a5..b26c50487351ad 100644 --- a/tests/sentry/snuba/metrics/test_metrics_layer/test_metrics_enhanced_performance.py +++ b/tests/sentry/snuba/metrics/test_metrics_layer/test_metrics_enhanced_performance.py @@ -27,11 +27,14 @@ MetricGroupByField, MetricOrderByField, MetricsQuery, +) +from sentry.snuba.metrics.datasource import get_custom_measurements, get_series +from sentry.snuba.metrics.naming_layer import ( + TransactionMetricKey, + TransactionMRI, TransactionStatusTagValue, TransactionTagsKey, ) -from sentry.snuba.metrics.datasource import get_custom_measurements, get_series -from sentry.snuba.metrics.naming_layer import TransactionMetricKey, TransactionMRI from sentry.snuba.metrics.query_builder import QueryDefinition from sentry.testutils import TestCase from sentry.testutils.cases import BaseMetricsLayerTestCase, MetricsEnhancedPerformanceTestCase @@ -2005,6 +2008,7 @@ def test_limit_when_not_passed_and_interval_is_provided(self): INTERVAL_LEN = 7 # 6 hours unaligned generate 7 1h intervals EXPECTED_DEFAULT_LIMIT = MAX_POINTS // INTERVAL_LEN + assert metrics_query.limit is not None assert metrics_query.limit.limit == EXPECTED_DEFAULT_LIMIT def test_high_limit_provided_not_raise_exception_when_high_interval_provided(self): @@ -2036,6 +2040,7 @@ def test_high_limit_provided_not_raise_exception_when_high_interval_provided(sel MetricsQuery(**metrics_query_dict) mq = MetricsQuery(**metrics_query_dict, interval=3600) + assert mq.limit is not None assert mq.limit.limit == 50 diff --git a/tests/snuba/search/test_backend.py b/tests/snuba/search/test_backend.py index b5a4240ba37bcb..694e724d296f15 100644 --- a/tests/snuba/search/test_backend.py +++ b/tests/snuba/search/test_backend.py @@ -1,5 +1,6 @@ import uuid from datetime import datetime, timedelta +from typing import Any from unittest import mock import pytest @@ -34,8 +35,10 @@ from sentry.search.snuba.backend import ( CdcEventsDatasetSnubaSearchBackend, EventsDatasetSnubaSearchBackend, + SnubaSearchBackendBase, ) from sentry.search.snuba.executors import InvalidQueryForExecutor, PrioritySortWeights +from sentry.snuba.dataset import Dataset from sentry.testutils import SnubaTestCase, TestCase, xfail_if_not_postgres from sentry.testutils.helpers import Feature from sentry.testutils.helpers.datetime import before_now, iso_format @@ -49,6 +52,10 @@ def date_to_query_format(date): class SharedSnubaTest(TestCase, SnubaTestCase): + @property + def backend(self) -> SnubaSearchBackendBase: + raise NotImplementedError(self) + def build_search_filter(self, query, projects=None, user=None, environments=None): user = user if user is not None else self.user projects = projects if projects is not None else [self.project] @@ -812,6 +819,7 @@ def test_age_filter_with_environment(self): group1_first_seen = GroupEnvironment.objects.get( environment=self.environments["production"], group=self.group1 ).first_seen + assert group1_first_seen is not None results = self.make_query( environments=[self.environments["production"]], @@ -2406,7 +2414,7 @@ def test_query(query): test_query("has:%s" % key) test_query("!has:%s" % key) if key == "error.handled": - val = 1 + val: Any = 1 elif key in issue_search_config.numeric_keys: val = "123" elif key in issue_search_config.date_keys: @@ -2926,6 +2934,7 @@ def test_priority_mixed_group_types(self): "received": before_now(minutes=1).isoformat(), }, ) + assert group_info is not None profile_group_1 = group_info.group agg_kwargs = { @@ -3215,7 +3224,7 @@ def test_perf_issue_search_message_term_queries_postgres(self): ] ): result = snuba.raw_query( - dataset=snuba.Dataset.IssuePlatform, + dataset=Dataset.IssuePlatform, start=self.base_datetime - timedelta(hours=1), end=self.base_datetime + timedelta(hours=1), selected_columns=[ @@ -3356,6 +3365,7 @@ def setUp(self): "received": before_now(minutes=1).isoformat(), }, ) + assert group_info is not None self.profile_group_1 = group_info.group event_id_2 = uuid.uuid4().hex @@ -3375,6 +3385,7 @@ def setUp(self): "received": before_now(minutes=2).isoformat(), }, ) + assert group_info is not None self.profile_group_2 = group_info.group event_id_3 = uuid.uuid4().hex @@ -3469,6 +3480,7 @@ def test_generic_query_perf(self): "received": before_now(minutes=1).isoformat(), }, ) + assert group_info is not None results = self.make_query(search_filter_query="issue.category:performance my_tag:2") assert list(results) == []
c8580f4bfed985a3586fe0f557c97f6e84bf2ff5
2019-02-25 19:50:30
Armin Ronacher
feat: Improved event titles for native crashes (#12131)
false
Improved event titles for native crashes (#12131)
feat
diff --git a/requirements-base.txt b/requirements-base.txt index 8d473a2124fca4..41f3b7e220e0ca 100644 --- a/requirements-base.txt +++ b/requirements-base.txt @@ -55,7 +55,7 @@ redis>=2.10.3,<2.10.6 requests-oauthlib==0.3.3 requests[security]>=2.20.0,<2.21.0 selenium==3.11.0 -semaphore>=0.4.15,<0.5.0 +semaphore>=0.4.16,<0.5.0 sentry-sdk>=0.7.0 setproctitle>=1.1.7,<1.2.0 simplejson>=3.2.0,<3.9.0 diff --git a/src/sentry/culprit.py b/src/sentry/culprit.py new file mode 100644 index 00000000000000..536427071e3485 --- /dev/null +++ b/src/sentry/culprit.py @@ -0,0 +1,72 @@ +""" +This file implements the legacy culprit system. The culprit at this point is +just used as a fallback if no transaction is set. When a transaction is set +the culprit is overriden by the transaction value. + +Over time we want to fully phase out the culprit. Until then this is the +code that generates it. +""" + +from __future__ import absolute_import +from sentry.constants import MAX_CULPRIT_LENGTH +from sentry.utils.safe import get_path +from sentry.utils.strings import truncatechars + + +def generate_culprit(data): + platform = data.get('platform') + exceptions = get_path(data, 'exception', 'values', filter=True) + if exceptions: + # Synthetic events no longer get a culprit + last_exception = get_path(exceptions, -1) + if get_path(last_exception, 'mechanism', 'synthetic'): + return '' + + stacktraces = [e['stacktrace'] for e in exceptions if get_path(e, 'stacktrace', 'frames')] + else: + stacktrace = data.get('stacktrace') + if stacktrace and stacktrace.get('frames'): + stacktraces = [stacktrace] + else: + stacktraces = None + + culprit = None + + if not culprit and stacktraces: + culprit = get_stacktrace_culprit(get_path(stacktraces, -1), platform=platform) + + if not culprit and data.get('request'): + culprit = get_path(data, 'request', 'url') + + return truncatechars(culprit or '', MAX_CULPRIT_LENGTH) + + +def get_stacktrace_culprit(stacktrace, platform): + default = None + for frame in reversed(stacktrace['frames']): + if not frame: + continue + if frame.get('in_app'): + culprit = get_frame_culprit(frame, platform=platform) + if culprit: + return culprit + elif default is None: + default = get_frame_culprit(frame, platform=platform) + return default + + +def get_frame_culprit(frame, platform): + # If this frame has a platform, we use it instead of the one that + # was passed in (as that one comes from the exception which might + # not necessarily be the same platform). + platform = frame.get('platform') or platform + if platform in ('objc', 'cocoa', 'native'): + return frame.get('function') or '?' + fileloc = frame.get('module') or frame.get('filename') + if not fileloc: + return '' + elif platform in ('javascript', 'node'): + # function and fileloc might be unicode here, so let it coerce + # to a unicode string if needed. + return '%s(%s)' % (frame.get('function') or '?', fileloc) + return '%s in %s' % (fileloc, frame.get('function') or '?') diff --git a/src/sentry/event_manager.py b/src/sentry/event_manager.py index b98f68afb40d8a..b7e5b09c29616a 100644 --- a/src/sentry/event_manager.py +++ b/src/sentry/event_manager.py @@ -18,8 +18,7 @@ from sentry import buffer, eventtypes, eventstream, features, tagstore, tsdb, filters from sentry.constants import ( - LOG_LEVELS, LOG_LEVELS_MAP, MAX_CULPRIT_LENGTH, VALID_PLATFORMS, - MAX_TAG_VALUE_LENGTH, + LOG_LEVELS, LOG_LEVELS_MAP, VALID_PLATFORMS, MAX_TAG_VALUE_LENGTH, ) from sentry.coreapi import ( APIError, @@ -52,11 +51,11 @@ from sentry.utils.dates import to_timestamp from sentry.utils.db import is_postgres, is_mysql from sentry.utils.safe import safe_execute, trim, get_path, setdefault_path -from sentry.utils.strings import truncatechars from sentry.utils.geo import rust_geoip from sentry.utils.validators import is_float from sentry.utils.contexts_normalization import normalize_user_agent from sentry.stacktraces import normalize_in_app +from sentry.culprit import generate_culprit logger = logging.getLogger("sentry.events") @@ -135,31 +134,6 @@ def should_sample(current_datetime, last_seen, times_seen): return True -def generate_culprit(data, platform=None): - exceptions = get_path(data, 'exception', 'values') - if exceptions: - stacktraces = [e['stacktrace'] for e in exceptions if get_path(e, 'stacktrace', 'frames')] - else: - stacktrace = data.get('stacktrace') - if stacktrace and stacktrace.get('frames'): - stacktraces = [stacktrace] - else: - stacktraces = None - - culprit = None - - if not culprit and stacktraces: - from sentry.interfaces.stacktrace import Stacktrace - culprit = Stacktrace.to_python(stacktraces[-1]).get_culprit_string( - platform=platform, - ) - - if not culprit and data.get('request'): - culprit = get_path(data, 'request', 'url') - - return truncatechars(culprit or '', MAX_CULPRIT_LENGTH) - - def plugin_is_regression(group, event): project = event.project for plugin in plugins.for_project(project): @@ -545,7 +519,7 @@ def get_culprit(self): return force_text( self._data.get('culprit') or self._data.get('transaction') or - generate_culprit(self._data, platform=self._data['platform']) or + generate_culprit(self._data) or '' ) diff --git a/src/sentry/eventtypes/error.py b/src/sentry/eventtypes/error.py index 50c3c3c4c46140..7c1a378196b095 100644 --- a/src/sentry/eventtypes/error.py +++ b/src/sentry/eventtypes/error.py @@ -8,15 +8,19 @@ from .base import BaseEvent -def get_crash_file(stacktrace): +def get_crash_location(exception, platform=None): default = None - for frame in reversed(get_path(stacktrace, 'frames', filter=True) or ()): + for frame in reversed(get_path(exception, 'stacktrace', 'frames', filter=True) or ()): fn = frame.get('filename') or frame.get('abs_path') if fn: + func = frame.get('function') + if func is not None: + from sentry.interfaces.stacktrace import trim_function_name + func = trim_function_name(func, frame.get('platform') or platform) if frame.get('in_app'): - return fn + return fn, func if default is None: - default = fn + default = fn, func return default @@ -29,28 +33,38 @@ def has_metadata(self, data): def get_metadata(self, data): exception = get_path(data, 'exception', 'values', -1) + if not exception: + return {} - # in some situations clients are submitting non-string data for these - rv = { - 'type': trim(get_path(exception, 'type', default='Error'), 128), - 'value': trim(get_path(exception, 'value', default=''), 1024), - } + loc = get_crash_location(exception, data.get('platform')) + rv = {} - # Attach crash location - if exception: - stacktrace = exception.get('stacktrace') - if stacktrace: - fn = get_crash_file(stacktrace) - if fn is not None: - rv['filename'] = fn + # If the exception mechanism indicates a synthetic exception we do not + # want to record the type and value into the metadata. + if not get_path(exception, 'mechanism', 'synthetic'): + rv.update({ + 'type': trim(get_path(exception, 'type', default='Error'), 128), + 'value': trim(get_path(exception, 'value', default=''), 1024), + }) + + # Attach crash location if available + if loc is not None: + fn, func = loc + if fn: + rv['filename'] = fn + if func: + rv['function'] = func return rv def get_title(self, metadata): + ty = metadata.get('type') + if ty is None: + return metadata.get('function') or '<unknown>' if not metadata['value']: - return metadata['type'] + return ty return u'{}: {}'.format( - metadata['type'], + ty, truncatechars(metadata['value'].splitlines()[0], 100), ) diff --git a/src/sentry/interfaces/exception.py b/src/sentry/interfaces/exception.py index 7cf56e76dd930d..2c6d303d58db83 100644 --- a/src/sentry/interfaces/exception.py +++ b/src/sentry/interfaces/exception.py @@ -710,6 +710,7 @@ class Mechanism(Interface): >>> "relevant_address": "0x1" >>> }, >>> "handled": false, + >>> "synthetic": false, >>> "help_link": "https://developer.apple.com/library/content/qa/qa1367/_index.html", >>> "meta": { >>> "mach_exception": { @@ -762,6 +763,7 @@ def to_python(cls, data): kwargs = { 'type': trim(data['type'], 128), + 'synthetic': data.get('synthetic'), 'description': trim(data.get('description'), 1024), 'help_link': trim(data.get('help_link'), 1024), 'handled': data.get('handled'), @@ -778,6 +780,7 @@ def to_python(cls, data): def to_json(self): return prune_empty_keys({ 'type': self.type, + 'synthetic': self.synthetic, 'description': self.description, 'help_link': self.help_link, 'handled': self.handled, diff --git a/src/sentry/interfaces/stacktrace.py b/src/sentry/interfaces/stacktrace.py index a44700a5742ce4..18a252cdc57b23 100644 --- a/src/sentry/interfaces/stacktrace.py +++ b/src/sentry/interfaces/stacktrace.py @@ -36,6 +36,10 @@ )/""", re.X | re.I ) +# Native function trim re. For now this is a simple hack until we have the +# language hints in which will let us trim this down better. +_native_function_trim_re = re.compile(r'^(.[^(]*)\(') + # OpenJDK auto-generated classes for reflection access: # sun.reflect.GeneratedSerializationConstructorAccessor123 # sun.reflect.GeneratedConstructorAccessor456 @@ -97,6 +101,20 @@ def trim_package(pkg): return pkg +def trim_function_name(func, platform): + # TODO(mitsuhiko): we actually want to use the language information here + # but we don't have that yet. + if platform in ('objc', 'cocoa', 'native'): + # objc function + if func.startswith(('[', '+[', '-[')): + return func + # c/c++ function hopefully + match = _native_function_trim_re.match(func.strip()) + if match is not None: + return match.group(1).strip() + return func + + def to_hex_addr(addr): if addr is None: return None @@ -622,23 +640,6 @@ def to_string(self, event): } ).strip('\n') - def get_culprit_string(self, platform=None): - # If this frame has a platform, we use it instead of the one that - # was passed in (as that one comes from the exception which might - # not necessarily be the same platform). - if self.platform is not None: - platform = self.platform - if platform in ('objc', 'cocoa', 'native'): - return self.function or '?' - fileloc = self.module or self.filename - if not fileloc: - return '' - elif platform in ('javascript', 'node'): - # function and fileloc might be unicode here, so let it coerce - # to a unicode string if needed. - return '%s(%s)' % (self.function or '?', fileloc) - return '%s in %s' % (fileloc, self.function or '?', ) - class Stacktrace(Interface): """ @@ -944,14 +945,3 @@ def get_stacktrace( ) return '\n'.join(result) - - def get_culprit_string(self, platform=None): - default = None - for frame in reversed(self.frames): - if frame.in_app: - culprit = frame.get_culprit_string(platform=platform) - if culprit: - return culprit - elif default is None: - default = frame.get_culprit_string(platform=platform) - return default diff --git a/src/sentry/lang/native/minidump.py b/src/sentry/lang/native/minidump.py index 9021ec18959567..2b3c2519308098 100644 --- a/src/sentry/lang/native/minidump.py +++ b/src/sentry/lang/native/minidump.py @@ -78,6 +78,7 @@ def merge_process_state_event(data, state, cfi=None): 'mechanism': { 'type': 'minidump', 'handled': False, + 'synthetic': True, # We cannot extract exception codes or signals with the breakpad # extractor just yet. Once these capabilities are added to symbolic, # these values should go in the mechanism here. diff --git a/src/sentry/static/sentry/app/components/eventOrGroupHeader.jsx b/src/sentry/static/sentry/app/components/eventOrGroupHeader.jsx index 67aa2ff271968c..c9f3cea4053ee6 100644 --- a/src/sentry/static/sentry/app/components/eventOrGroupHeader.jsx +++ b/src/sentry/static/sentry/app/components/eventOrGroupHeader.jsx @@ -63,8 +63,11 @@ class EventOrGroupHeader extends React.Component { getLocation() { const {data} = this.props; - const {metadata} = data || {}; - return metadata.filename || null; + if (data.type === 'error') { + const {metadata} = data || {}; + return metadata.filename || null; + } + return null; } getTitle() { diff --git a/src/sentry/static/sentry/app/components/eventOrGroupTitle.jsx b/src/sentry/static/sentry/app/components/eventOrGroupTitle.jsx index 2cd36eadacc06f..a8807b82550bf7 100644 --- a/src/sentry/static/sentry/app/components/eventOrGroupTitle.jsx +++ b/src/sentry/static/sentry/app/components/eventOrGroupTitle.jsx @@ -26,8 +26,12 @@ class EventOrGroupTitle extends React.Component { let subtitle = null; if (type == 'error') { - title = metadata.type; subtitle = culprit; + if (metadata.type) { + title = metadata.type; + } else { + title = metadata.function || '<unknown>'; + } } else if (type == 'csp') { title = metadata.directive; subtitle = metadata.uri; diff --git a/src/sentry/tasks/unmerge.py b/src/sentry/tasks/unmerge.py index 4904df54a5d37e..03672150b71520 100644 --- a/src/sentry/tasks/unmerge.py +++ b/src/sentry/tasks/unmerge.py @@ -75,11 +75,19 @@ def merge_mappings(values): return result +def _generate_culprit(event): + # XXX(mitsuhiko): workaround: some old events do not have this data yet. + # This should be save delete by end of 2019 even considering slow on-prem + # releases. Platform was added back to data in december 2018. + data = event.data + if data.get('platform') is None: + data = dict(data.items()) + data['platform'] = event.platform + return generate_culprit(data) + + initial_fields = { - 'culprit': lambda event: generate_culprit( - event.data, - event.platform, - ), + 'culprit': lambda event: _generate_culprit(event), 'data': lambda event: { 'last_received': event.data.get('received') or float(event.datetime.strftime('%s')), 'type': event.data['type'], diff --git a/tests/sentry/interfaces/test_stacktrace.py b/tests/sentry/interfaces/test_stacktrace.py index 47c6eb2e7603a6..5e92b5a09f0564 100644 --- a/tests/sentry/interfaces/test_stacktrace.py +++ b/tests/sentry/interfaces/test_stacktrace.py @@ -9,7 +9,10 @@ from django.template.loader import render_to_string from exam import fixture -from sentry.interfaces.stacktrace import (Frame, Stacktrace, get_context, is_url, slim_frame_data) +from sentry.interfaces.stacktrace import ( + Frame, Stacktrace, get_context, is_url, slim_frame_data, + trim_function_name +) from sentry.models import Event from sentry.testutils import TestCase @@ -25,6 +28,16 @@ def test_is_url(): assert is_url('blob:\x00') is False +def test_trim_function_name(): + assert trim_function_name('+[foo:(bar)]', 'objc') == '+[foo:(bar)]' + assert trim_function_name('[foo:(bar)]', 'objc') == '[foo:(bar)]' + assert trim_function_name('-[foo:(bar)]', 'objc') == '-[foo:(bar)]' + assert trim_function_name( + '(anonymous namespace)::foo(int)', + 'native') == '(anonymous namespace)::foo' + assert trim_function_name('foo::bar::foo(int)', 'native') == 'foo::bar::foo' + + class GetContextTest(TestCase): def test_works_with_empty_filename(self): result = get_context(0, 'hello world') @@ -721,64 +734,6 @@ def test_compute_hashes_ignores_safari_native_code(self): result = interface.compute_hashes() self.assertEquals(result, []) - def test_cocoa_culprit(self): - stacktrace = Stacktrace.to_python( - dict( - frames=[ - { - 'filename': 'foo/baz.c', - 'package': '/foo/bar/baz.dylib', - 'lineno': 1, - 'in_app': True, - 'function': '-[CRLCrashAsyncSafeThread crash]', - } - ] - ) - ) - assert stacktrace.get_culprit_string(platform='cocoa') == '-[CRLCrashAsyncSafeThread crash]' - - def test_emoji_culprit(self): - stacktrace = Stacktrace.to_python( - dict( - frames=[ - { - 'filename': 'foo/baz.c', - 'package': '/foo/bar/baz.dylib', - 'module': u'\U0001f62d', - 'lineno': 1, - 'in_app': True, - 'function': u'\U0001f60d', - } - ] - ) - ) - assert stacktrace.get_culprit_string(platform='javascript') == u'\U0001f60d(\U0001f62d)' - - def test_cocoa_strict_stacktrace(self): - stacktrace = Stacktrace.to_python( - dict( - frames=[ - { - 'filename': 'foo/baz.c', - 'package': '/foo/bar/libswiftCore.dylib', - 'lineno': 1, - 'in_app': False, - 'function': 'fooBar', - }, { - 'package': '/foo/bar/MyApp', - 'in_app': True, - 'function': 'fooBar2', - }, { - 'filename': 'Mycontroller.swift', - 'package': '/foo/bar/MyApp', - 'in_app': True, - 'function': '-[CRLCrashAsyncSafeThread crash]', - } - ] - ) - ) - assert stacktrace.get_culprit_string(platform='cocoa') == '-[CRLCrashAsyncSafeThread crash]' - def test_compute_hashes_does_not_group_different_js_errors(self): interface = Stacktrace.to_python( { diff --git a/tests/sentry/lang/javascript/test_processor.py b/tests/sentry/lang/javascript/test_processor.py index 2c58edfa363a01..6d8238c772494a 100644 --- a/tests/sentry/lang/javascript/test_processor.py +++ b/tests/sentry/lang/javascript/test_processor.py @@ -569,7 +569,7 @@ def test_get_culprit_is_patched(self): } generate_modules(data) fix_culprit(data) - assert data['culprit'] == 'bar in oops' + assert data['culprit'] == 'oops(bar)' def test_ensure_module_names(self): from sentry.lang.javascript.plugin import generate_modules diff --git a/tests/sentry/lang/native/test_minidump.py b/tests/sentry/lang/native/test_minidump.py index f51da3cf03d70e..3d0116c8758569 100644 --- a/tests/sentry/lang/native/test_minidump.py +++ b/tests/sentry/lang/native/test_minidump.py @@ -148,7 +148,8 @@ def test_minidump_linux(): 'exception': { 'mechanism': { 'type': 'minidump', - 'handled': False + 'handled': False, + 'synthetic': True, }, 'stacktrace': { 'frames': [ @@ -625,7 +626,8 @@ def test_minidump_macos(): 'exception': { 'mechanism': { 'type': 'minidump', - 'handled': False + 'handled': False, + 'synthetic': True, }, 'stacktrace': { 'frames': [ @@ -836,7 +838,8 @@ def test_minidump_windows(): 'exception': { 'mechanism': { 'type': 'minidump', - 'handled': False + 'handled': False, + 'synthetic': True, }, 'stacktrace': { 'frames': [ diff --git a/tests/sentry/test_culprit.py b/tests/sentry/test_culprit.py new file mode 100644 index 00000000000000..73c079c6ee6719 --- /dev/null +++ b/tests/sentry/test_culprit.py @@ -0,0 +1,154 @@ +from __future__ import absolute_import + +from sentry.event_manager import EventManager + + +def get_culprit(data): + mgr = EventManager(data) + mgr.normalize() + return mgr.get_culprit() + + +def test_cocoa_culprit(): + culprit = get_culprit({ + 'platform': 'cocoa', + 'exception': { + 'type': 'Crash', + 'stacktrace': { + 'frames': [ + { + 'filename': 'foo/baz.c', + 'package': '/foo/bar/baz.dylib', + 'lineno': 1, + 'in_app': True, + 'function': '-[CRLCrashAsyncSafeThread crash]', + } + ] + } + } + }) + assert culprit == '-[CRLCrashAsyncSafeThread crash]' + + +def test_emoji_culprit(): + culprit = get_culprit({ + 'platform': 'native', + 'exception': { + 'type': 'Crash', + 'stacktrace': { + 'frames': [ + { + 'filename': 'foo/baz.c', + 'package': '/foo/bar/baz.dylib', + 'module': u'\U0001f62d', + 'lineno': 1, + 'in_app': True, + 'function': u'\U0001f60d', + } + ] + } + } + }) + assert culprit == u'\U0001f60d' + + +def test_cocoa_strict_stacktrace(): + culprit = get_culprit({ + 'platform': 'native', + 'exception': { + 'type': 'Crash', + 'stacktrace': { + 'frames': [ + { + 'filename': 'foo/baz.c', + 'package': '/foo/bar/libswiftCore.dylib', + 'lineno': 1, + 'in_app': False, + 'function': 'fooBar', + }, { + 'package': '/foo/bar/MyApp', + 'in_app': True, + 'function': 'fooBar2', + }, { + 'filename': 'Mycontroller.swift', + 'package': '/foo/bar/MyApp', + 'in_app': True, + 'function': '-[CRLCrashAsyncSafeThread crash]', + } + ] + } + } + }) + assert culprit == '-[CRLCrashAsyncSafeThread crash]' + + +def test_culprit_for_synthetic_event(): + # Synthetic events do not generate a culprit + culprit = get_culprit({ + 'platform': 'javascript', + 'exception': { + 'type': 'Error', + 'value': 'I threw up stringly', + 'mechanism': { + 'type': 'string-error', + 'synthetic': True, + }, + 'stacktrace': { + 'frames': [ + { + 'filename': 'foo/baz.js', + 'package': 'node_modules/blah/foo/bar.js', + 'lineno': 42, + 'in_app': True, + 'function': 'fooBar', + } + ] + } + } + }) + assert culprit == '' + + +def test_culprit_for_javascript_event(): + culprit = get_culprit({ + 'platform': 'javascript', + 'exception': { + 'type': 'Error', + 'value': 'I fail hard', + 'stacktrace': { + 'frames': [ + { + 'filename': 'foo/baz.js', + 'package': 'node_modules/blah/foo/bar.js', + 'lineno': 42, + 'in_app': True, + 'function': 'fooBar', + } + ] + } + } + }) + assert culprit == 'fooBar(foo/baz.js)' + + +def test_culprit_for_python_event(): + culprit = get_culprit({ + 'platform': 'python', + 'exception': { + 'type': 'ZeroDivisionError', + 'value': 'integer division or modulo by zero', + 'stacktrace': { + 'frames': [ + { + 'filename': 'foo/baz.py', + 'module': 'foo.baz', + 'package': 'foo/baz.py', + 'lineno': 23, + 'in_app': True, + 'function': 'it_failed', + } + ] + } + } + }) + assert culprit == 'foo.baz in it_failed'
5f349c500cdc78abd4fb4dae07adb8d4add669e6
2022-07-07 18:15:54
Floris Bruynooghe
ref(projectconfig): Give build task more time (#36417)
false
Give build task more time (#36417)
ref
diff --git a/src/sentry/tasks/relay.py b/src/sentry/tasks/relay.py index 0150c62c076a5a..bc580d47da9f5b 100644 --- a/src/sentry/tasks/relay.py +++ b/src/sentry/tasks/relay.py @@ -19,8 +19,8 @@ name="sentry.tasks.relay.build_project_config", queue="relay_config", acks_late=True, - soft_time_limit=5, - time_limit=10, # Extra 5 seconds to remove the debounce key + soft_time_limit=25, + time_limit=30, # Extra 5 seconds to remove the debounce key ) def build_project_config(public_key=None, **kwargs): """Build a project config and put it in the Redis cache.
2786b197c04fd53bb73117a8938306e7774e8d8f
2022-08-29 20:57:11
Evan Purkhiser
ref(js): Convert FormPanel to a FC (#38157)
false
Convert FormPanel to a FC (#38157)
ref
diff --git a/static/app/components/forms/formPanel.tsx b/static/app/components/forms/formPanel.tsx index 314024be3bed68..e284dc40214a7d 100644 --- a/static/app/components/forms/formPanel.tsx +++ b/static/app/components/forms/formPanel.tsx @@ -1,4 +1,4 @@ -import {Component} from 'react'; +import {useCallback, useState} from 'react'; import styled from '@emotion/styled'; import FieldFromConfig from 'sentry/components/forms/fieldFromConfig'; @@ -9,131 +9,105 @@ import {sanitizeQuerySelector} from 'sentry/utils/sanitizeQuerySelector'; import {FieldObject, JsonFormObject} from './type'; -type DefaultProps = { - additionalFieldProps: {[key: string]: any}; -}; - -type Props = DefaultProps & { +type Props = { /** * List of fields to render */ fields: FieldObject[]; - access?: Set<Scope>; - - /** Can the PanelBody be hidden with a click? */ + additionalFieldProps?: {[key: string]: any}; + /** + * Can the PanelBody be hidden with a click? + */ collapsible?: boolean; /** * Disables the entire form */ disabled?: boolean; - features?: Record<string, any>; - /** * The name of the field that should be highlighted */ highlighted?: string; - /** * Renders inside of PanelBody before PanelBody close */ renderFooter?: (arg: JsonFormObject) => React.ReactNode; - /** * Renders inside of PanelBody at the start */ renderHeader?: (arg: JsonFormObject) => React.ReactNode; - /** * Panel title */ title?: React.ReactNode; }; -type State = { - collapsed: boolean; -}; - -export default class FormPanel extends Component<Props, State> { - static defaultProps: DefaultProps = { - additionalFieldProps: {}, - }; - - state: State = { - collapsed: false, - }; - - handleToggleEvents = () => { - const {collapsed} = this.state; - - this.setState({collapsed: !collapsed}); - }; - - render() { - const { - title, - fields, - access, - disabled, - additionalFieldProps, - renderFooter, - renderHeader, - collapsible, - ...otherProps - } = this.props; - const {collapsed} = this.state; - - return ( - <Panel id={typeof title === 'string' ? sanitizeQuerySelector(title) : undefined}> - {title && ( - <PanelHeader> - {title} - {collapsible && ( - <Collapse onClick={this.handleToggleEvents}> - <IconChevron direction={collapsed ? 'down' : 'up'} size="xs" /> - </Collapse> - )} - </PanelHeader> - )} - {!collapsed && ( - <PanelBody> - {typeof renderHeader === 'function' && renderHeader({title, fields})} - - {fields.map(field => { - if (typeof field === 'function') { - return field(); - } - - const {defaultValue: _, ...fieldWithoutDefaultValue} = field; - - // Allow the form panel disabled prop to override the fields - // disabled prop, with fallback to the fields disabled state. - if (disabled === true) { - fieldWithoutDefaultValue.disabled = true; - fieldWithoutDefaultValue.disabledReason = undefined; - } - - return ( - <FieldFromConfig - access={access} - disabled={disabled} - key={field.name} - {...otherProps} - {...additionalFieldProps} - field={fieldWithoutDefaultValue} - highlighted={this.props.highlighted === `#${field.name}`} - /> - ); - })} - {typeof renderFooter === 'function' && renderFooter({title, fields})} - </PanelBody> - )} - </Panel> - ); - } +function FormPanel({ + additionalFieldProps = {}, + title, + fields, + access, + disabled, + renderFooter, + renderHeader, + collapsible, + ...otherProps +}: Props) { + const [collapsed, setCollapse] = useState(false); + const handleCollapseToggle = useCallback(() => setCollapse(current => !current), []); + + return ( + <Panel id={typeof title === 'string' ? sanitizeQuerySelector(title) : undefined}> + {title && ( + <PanelHeader> + {title} + {collapsible && ( + <Collapse onClick={handleCollapseToggle}> + <IconChevron direction={collapsed ? 'down' : 'up'} size="xs" /> + </Collapse> + )} + </PanelHeader> + )} + {!collapsed && ( + <PanelBody> + {typeof renderHeader === 'function' && renderHeader({title, fields})} + + {fields.map(field => { + if (typeof field === 'function') { + return field(); + } + + const {defaultValue: _, ...fieldWithoutDefaultValue} = field; + + // Allow the form panel disabled prop to override the fields + // disabled prop, with fallback to the fields disabled state. + if (disabled === true) { + fieldWithoutDefaultValue.disabled = true; + fieldWithoutDefaultValue.disabledReason = undefined; + } + + return ( + <FieldFromConfig + access={access} + disabled={disabled} + key={field.name} + {...otherProps} + {...additionalFieldProps} + field={fieldWithoutDefaultValue} + highlighted={otherProps.highlighted === `#${field.name}`} + /> + ); + })} + {typeof renderFooter === 'function' && renderFooter({title, fields})} + </PanelBody> + )} + </Panel> + ); } +export default FormPanel; + const Collapse = styled('span')` cursor: pointer; `;
4fbbde7a4239775f6ec5f3c247cd2e5522eadfbf
2024-11-20 01:43:35
Matthew T
fix(security): ensure user views are isolated (#80994)
false
ensure user views are isolated (#80994)
fix
diff --git a/src/sentry/issues/endpoints/organization_group_search_views.py b/src/sentry/issues/endpoints/organization_group_search_views.py index e0f9f123bc2764..99d46dab86c745 100644 --- a/src/sentry/issues/endpoints/organization_group_search_views.py +++ b/src/sentry/issues/endpoints/organization_group_search_views.py @@ -118,7 +118,6 @@ def put(self, request: Request, organization: Organization) -> Response: def bulk_update_views( org: Organization, user_id: int, views: list[GroupSearchViewValidatorResponse] ) -> None: - existing_view_ids = [view["id"] for view in views if "id" in view] _delete_missing_views(org, user_id, view_ids_to_keep=existing_view_ids) @@ -140,7 +139,7 @@ def _update_existing_view( org: Organization, user_id: int, view: GroupSearchViewValidatorResponse, position: int ) -> None: try: - GroupSearchView.objects.get(id=view["id"]).update( + GroupSearchView.objects.get(id=view["id"], user_id=user_id).update( name=view["name"], query=view["query"], query_sort=view["querySort"], diff --git a/tests/sentry/issues/endpoints/test_organization_group_search_views.py b/tests/sentry/issues/endpoints/test_organization_group_search_views.py index c08f0382211982..f65aba051fc6b4 100644 --- a/tests/sentry/issues/endpoints/test_organization_group_search_views.py +++ b/tests/sentry/issues/endpoints/test_organization_group_search_views.py @@ -332,3 +332,80 @@ def test_updated_deleted_view(self) -> None: assert response.data[1]["query"] == view_one["query"] assert response.data[1]["querySort"] == view_one["querySort"] assert response.data[2] == views[2] + + +class OrganizationGroupSearchViewsPutRegressionTest(APITestCase): + endpoint = "sentry-api-0-organization-group-search-views" + method = "put" + + def setUp(self) -> None: + self.user_2 = self.create_user() + self.create_member(organization=self.organization, user=self.user_2) + + self.url = reverse( + "sentry-api-0-organization-group-search-views", + kwargs={"organization_id_or_slug": self.organization.slug}, + ) + + @with_feature({"organizations:issue-stream-custom-views": True}) + def test_cannot_rename_other_users_views(self) -> None: + self.login_as(user=self.user) + views = self.client.get(self.url).data + view = views[0] + + # ensure we only have the default view + assert len(views) == 1 + assert view["name"] == "Prioritized" + assert view["query"] == "is:unresolved issue.priority:[high, medium]" + assert view["querySort"] == "date" + assert view["position"] == 0 + + # create a new custom view + views.append( + { + "name": "Custom View Two", + "query": "is:unresolved", + "query_sort": "date", + } + ) + + response = self.get_success_response(self.organization.slug, views=views) + + assert len(response.data) == 2 # 1 existing default view + 1 new view + assert response.data[1]["name"] == "Custom View Two" + assert response.data[1]["query"] == "is:unresolved" + assert response.data[1]["querySort"] == "date" + + # now "delete" the custom view so the default view gets a proper ID + views = self.client.get(self.url).data + views.pop(1) + + response = self.get_success_response(self.organization.slug, views=views) + + # we should only have the default view now + assert len(response.data) == 1 + assert response.data[0]["name"] == "Prioritized" + assert response.data[0]["id"] # and it should now have an ID + + # attempt to change user's 1 view from user 2 + views = self.client.get(self.url).data + default_view = views[0] + default_view["name"] = "New Name" + + self.login_as(user=self.user_2) + response = self.get_success_response(self.organization.slug, views=views) + + # instead of editing the original view, it should create a new view for user 2 + assert len(response.data) == 1 + assert response.data[0]["id"] != default_view["id"] + assert response.data[0]["name"] == "New Name" + + # as user 1, verify the name has not been changed + self.login_as(user=self.user) + response = self.client.get(self.url) + + assert len(response.data) == 1 + assert response.data[0]["id"] == default_view["id"] + assert response.data[0]["name"] == "Prioritized" + assert response.data[0]["query"] == view["query"] + assert response.data[0]["querySort"] == view["querySort"]
034a5d96a0d4d8d955ee0e407556ecdf13d8d8fb
2023-11-01 07:33:19
Michelle Zhang
ref(feedback): when bulk resolve/unresolve, header checkbox unchecks (#59147)
false
when bulk resolve/unresolve, header checkbox unchecks (#59147)
ref
diff --git a/static/app/components/feedback/list/feedbackList.tsx b/static/app/components/feedback/list/feedbackList.tsx index bf625907c0c3b7..858c211f79c6a7 100644 --- a/static/app/components/feedback/list/feedbackList.tsx +++ b/static/app/components/feedback/list/feedbackList.tsx @@ -43,7 +43,7 @@ export default function FeedbackList() { const {setParamValue} = useUrlParams('query'); const clearSearchTerm = () => setParamValue(''); - const {checked, toggleChecked} = useListItemCheckboxState(); + const {checked, uncheckAll, toggleChecked} = useListItemCheckboxState(); const listRef = useRef<ReactVirtualizedList>(null); @@ -87,7 +87,7 @@ export default function FeedbackList() { return ( <Fragment> - <FeedbackListHeader checked={checked} toggleChecked={toggleChecked} /> + <FeedbackListHeader checked={checked} uncheckAll={uncheckAll} /> <OverflowPanelItem noPadding> <InfiniteLoader isRowLoaded={isRowLoaded} diff --git a/static/app/components/feedback/list/feedbackListHeader.tsx b/static/app/components/feedback/list/feedbackListHeader.tsx index d7cd0d4ac92f2b..b1fd49fe691f4c 100644 --- a/static/app/components/feedback/list/feedbackListHeader.tsx +++ b/static/app/components/feedback/list/feedbackListHeader.tsx @@ -24,10 +24,10 @@ import useUrlParams from 'sentry/utils/useUrlParams'; interface Props { checked: string[]; - toggleChecked: (id: string) => void; + uncheckAll: (ids: string[]) => void; } -export default function FeedbackListHeader({checked, toggleChecked}: Props) { +export default function FeedbackListHeader({checked, uncheckAll}: Props) { const {mailbox} = useLocationQuery({ fields: { mailbox: decodeMailbox, @@ -40,11 +40,11 @@ export default function FeedbackListHeader({checked, toggleChecked}: Props) { <Checkbox checked={checked.length ? 'indeterminate' : false} onChange={() => { - checked.length ? checked.forEach(c => toggleChecked(c)) : null; + checked.length ? uncheckAll(checked) : null; }} /> {checked.length ? ( - <HasSelection checked={checked} mailbox={mailbox} /> + <HasSelection checked={checked} mailbox={mailbox} uncheckAll={uncheckAll} /> ) : ( <MailboxPicker value={mailbox} onChange={setMailbox} /> )} @@ -52,7 +52,7 @@ export default function FeedbackListHeader({checked, toggleChecked}: Props) { ); } -function HasSelection({checked, mailbox}) { +function HasSelection({checked, mailbox, uncheckAll}) { const organization = useOrganization(); const {markAsRead, resolve} = useMutateFeedback({ feedbackIds: checked, @@ -81,6 +81,7 @@ function HasSelection({checked, mailbox}) { const newStatus = mailbox === 'resolved' ? GroupStatus.UNRESOLVED : GroupStatus.RESOLVED; resolve(newStatus, mutationOptions); + uncheckAll(checked); }} > {mailbox === 'resolved' ? t('Unresolve') : t('Resolve')} diff --git a/static/app/components/feedback/list/useListItemCheckboxState.tsx b/static/app/components/feedback/list/useListItemCheckboxState.tsx index ab9e07bdba5740..91c4e153e9c683 100644 --- a/static/app/components/feedback/list/useListItemCheckboxState.tsx +++ b/static/app/components/feedback/list/useListItemCheckboxState.tsx @@ -17,8 +17,13 @@ export default function useListItemCheckboxState() { }); }, []); + const uncheckAll = useCallback(() => { + setState({}); + }, []); + return { checked, toggleChecked, + uncheckAll, }; }
282520904348cb6a35230200978c7b69cb2fed85
2022-08-13 02:27:12
Gilbert Szeto
fix(releases): filter out releases with projects by project membership instead of access (#37495)
false
filter out releases with projects by project membership instead of access (#37495)
fix
diff --git a/src/sentry/api/bases/organization.py b/src/sentry/api/bases/organization.py index 7a1f147b7ea47b..d60ca824352030 100644 --- a/src/sentry/api/bases/organization.py +++ b/src/sentry/api/bases/organization.py @@ -379,7 +379,9 @@ def convert_args(self, request: Request, organization_slug=None, *args, **kwargs class OrganizationReleasesBaseEndpoint(OrganizationEndpoint): permission_classes = (OrganizationReleasePermission,) - def get_projects(self, request: Request, organization, project_ids=None): + def get_projects( + self, request: Request, organization, project_ids=None, include_all_accessible=True + ): """ Get all projects the current user or API token has access to. More detail in the parent class's method of the same name. @@ -401,7 +403,7 @@ def get_projects(self, request: Request, organization, project_ids=None): request, organization, force_global_perms=has_valid_api_key, - include_all_accessible=True, + include_all_accessible=include_all_accessible, project_ids=project_ids, ) diff --git a/src/sentry/api/endpoints/organization_releases.py b/src/sentry/api/endpoints/organization_releases.py index 1050db0568d4f2..9fbdfe8a94e97b 100644 --- a/src/sentry/api/endpoints/organization_releases.py +++ b/src/sentry/api/endpoints/organization_releases.py @@ -217,6 +217,14 @@ class OrganizationReleasesEndpoint( ] ) + def get_projects(self, request: Request, organization, project_ids=None): + return super().get_projects( + request, + organization, + project_ids=project_ids, + include_all_accessible="GET" != request.method, + ) + def get(self, request: Request, organization) -> Response: """ List an Organization's Releases diff --git a/tests/apidocs/endpoints/releases/test_organization_releases.py b/tests/apidocs/endpoints/releases/test_organization_releases.py index 7e46835ae60a80..f9fd0e225413a4 100644 --- a/tests/apidocs/endpoints/releases/test_organization_releases.py +++ b/tests/apidocs/endpoints/releases/test_organization_releases.py @@ -15,6 +15,8 @@ def setUp(self): team1 = self.create_team(organization=org) team2 = self.create_team(organization=org) + self.create_team_membership(team1, user=user) + self.create_team_membership(team2, user=user) self.project1 = self.create_project(teams=[team1], organization=org) self.project2 = self.create_project(teams=[team2], organization=org2) diff --git a/tests/sentry/api/endpoints/test_organization_member_team_details.py b/tests/sentry/api/endpoints/test_organization_member_team_details.py index 301c0ad8c71f9d..bddc9ea2f1fe82 100644 --- a/tests/sentry/api/endpoints/test_organization_member_team_details.py +++ b/tests/sentry/api/endpoints/test_organization_member_team_details.py @@ -1,6 +1,7 @@ from exam import fixture from rest_framework import status +from sentry.auth import access from sentry.models import ( Organization, OrganizationAccessRequest, @@ -442,6 +443,33 @@ def test_owner_can_remove_members(self): team=self.team, organizationmember=self.owner_on_team ).exists() + def test_access_revoked_after_leaving_team(self): + user = self.create_user() + organization = self.create_organization(flags=0) + team = self.create_team(organization=organization) + project = self.create_project(organization=organization, teams=[team]) + member = self.create_member(organization=organization, user=user, teams=[team]) + + ax = access.from_user(user, organization) + + # user a member of the team that is a part of the project should have the following access and scopes + assert ax.has_team_access(team) + assert ax.has_project_access(project) + assert ax.has_project_membership(project) + + self.login_as(user) + self.get_success_response( + organization.slug, member.id, team.slug, status_code=status.HTTP_200_OK + ) + + assert OrganizationMember.objects.filter(id=member.id).exists() + assert not OrganizationMemberTeam.objects.filter(organizationmember=member.id).exists() + + ax_after_leaving = access.from_user(user, organization) + assert not ax_after_leaving.has_team_access(team) + assert not ax_after_leaving.has_project_access(project) + assert not ax_after_leaving.has_project_membership(project) + class ReadOrganizationMemberTeamTest(OrganizationMemberTeamTestBase): endpoint = "sentry-api-0-organization-member-team-details" diff --git a/tests/sentry/api/endpoints/test_organization_releases.py b/tests/sentry/api/endpoints/test_organization_releases.py index 40add2ddedd20f..630e6b6c79b7ec 100644 --- a/tests/sentry/api/endpoints/test_organization_releases.py +++ b/tests/sentry/api/endpoints/test_organization_releases.py @@ -13,6 +13,7 @@ ReleaseSerializerWithProjects, ) from sentry.app import locks +from sentry.auth import access from sentry.constants import BAD_RELEASE_CHARS, MAX_COMMIT_LENGTH, MAX_VERSION_LENGTH from sentry.models import ( Activity, @@ -628,6 +629,52 @@ def test_project_permissions(self): ) release3.add_project(project1) + ax = access.from_user(user, org) + assert ax.has_projects_access([project1]) + assert ax.has_project_membership(project1) + assert not ax.has_project_membership(project2) + + response = self.get_success_response(org.slug) + self.assert_expected_versions(response, [release1, release3]) + + def test_project_permissions_open_access(self): + user = self.create_user(is_staff=False, is_superuser=False) + org = self.create_organization() + org.flags.allow_joinleave = True + org.save() + + team1 = self.create_team(organization=org) + team2 = self.create_team(organization=org) + + project1 = self.create_project(teams=[team1], organization=org) + project2 = self.create_project(teams=[team2], organization=org) + + self.create_member(teams=[team1], user=user, organization=org) + self.login_as(user=user) + + release1 = Release.objects.create( + organization_id=org.id, version="1", date_added=datetime(2013, 8, 13, 3, 8, 24, 880386) + ) + release1.add_project(project1) + + release2 = Release.objects.create( + organization_id=org.id, version="2", date_added=datetime(2013, 8, 14, 3, 8, 24, 880386) + ) + release2.add_project(project2) + + release3 = Release.objects.create( + organization_id=org.id, + version="3", + date_added=datetime(2013, 8, 12, 3, 8, 24, 880386), + date_released=datetime(2013, 8, 15, 3, 8, 24, 880386), + ) + release3.add_project(project1) + + ax = access.from_user(user, org) + assert ax.has_projects_access([project1, project2]) + assert ax.has_project_membership(project1) + assert not ax.has_project_membership(project2) + response = self.get_success_response(org.slug) self.assert_expected_versions(response, [release1, release3]) @@ -1794,7 +1841,7 @@ class OrganizationReleaseListEnvironmentsTest(APITestCase): def setUp(self): self.login_as(user=self.user) org = self.create_organization(owner=self.user) - team = self.create_team(organization=org) + team = self.create_team(organization=org, members=[self.user]) project1 = self.create_project(organization=org, teams=[team], name="foo") project2 = self.create_project(organization=org, teams=[team], name="bar")
f5eb3ec91f6e16b60943b63013a5f6299f967aa5
2023-02-27 23:01:52
Tony Xiao
feat(profiling): Use transactions for profiling search bar (#45105)
false
Use transactions for profiling search bar (#45105)
feat
diff --git a/static/app/utils/profiling/hooks/useProfileFilters.tsx b/static/app/utils/profiling/hooks/useProfileFilters.tsx index 14b98a870ae939..4e4045b7ee764d 100644 --- a/static/app/utils/profiling/hooks/useProfileFilters.tsx +++ b/static/app/utils/profiling/hooks/useProfileFilters.tsx @@ -8,17 +8,22 @@ import useOrganization from 'sentry/utils/useOrganization'; interface ProfileFiltersOptions { query: string; + disabled?: boolean; selection?: PageFilters; } -function useProfileFilters({query, selection}: ProfileFiltersOptions): TagCollection { +function useProfileFilters({ + query, + selection, + disabled, +}: ProfileFiltersOptions): TagCollection { const api = useApi(); const organization = useOrganization(); const [profileFilters, setProfileFilters] = useState<TagCollection>({}); useEffect(() => { - if (!selection) { + if (disabled || !selection) { return undefined; } @@ -44,7 +49,7 @@ function useProfileFilters({query, selection}: ProfileFiltersOptions): TagCollec }); return () => api.clear(); - }, [api, organization, query, selection]); + }, [api, organization, query, selection, disabled]); return profileFilters; } diff --git a/static/app/views/profiling/content.tsx b/static/app/views/profiling/content.tsx index 28fc14a7fd8c44..a1098206238524 100644 --- a/static/app/views/profiling/content.tsx +++ b/static/app/views/profiling/content.tsx @@ -8,6 +8,7 @@ import {Button} from 'sentry/components/button'; import ButtonBar from 'sentry/components/buttonBar'; import DatePageFilter from 'sentry/components/datePageFilter'; import EnvironmentPageFilter from 'sentry/components/environmentPageFilter'; +import SearchBar from 'sentry/components/events/searchBar'; import FeatureBadge from 'sentry/components/featureBadge'; import * as Layout from 'sentry/components/layouts/thirds'; import PageFilterBar from 'sentry/components/organizations/pageFilterBar'; @@ -25,6 +26,7 @@ import {t} from 'sentry/locale'; import SidebarPanelStore from 'sentry/stores/sidebarPanelStore'; import {space} from 'sentry/styles/space'; import trackAdvancedAnalyticsEvent from 'sentry/utils/analytics/trackAdvancedAnalyticsEvent'; +import EventView from 'sentry/utils/discover/eventView'; import { formatError, formatSort, @@ -55,7 +57,15 @@ function ProfilingContent({location}: ProfilingContentProps) { order: 'desc', }); - const profileFilters = useProfileFilters({query: '', selection}); + const profilingUsingTransactions = organization.features.includes( + 'profiling-using-transactions' + ); + + const profileFilters = useProfileFilters({ + query: '', + selection, + disabled: profilingUsingTransactions, + }); const {projects} = useProjects(); const transactions = useProfileEvents<FieldType>({ @@ -121,6 +131,22 @@ function ProfilingContent({location}: ProfilingContentProps) { 'profiling-dashboard-redesign' ); + const eventView = useMemo(() => { + const _eventView = EventView.fromNewQueryWithLocation( + { + id: undefined, + version: 2, + name: t('Profiling'), + fields: [], + query, + projects: selection.projects, + }, + location + ); + _eventView.additionalConditions.setFilterValues('has', ['profile.id']); + return _eventView; + }, [location, query, selection.projects]); + return ( <SentryDocumentTitle title={t('Profiling')} orgSlug={organization.slug}> <PageFiltersContainer> @@ -172,15 +198,26 @@ function ProfilingContent({location}: ProfilingContentProps) { <EnvironmentPageFilter /> <DatePageFilter alignDropdown="left" /> </PageFilterBar> - <SmartSearchBar - organization={organization} - hasRecentSearches - searchSource="profile_landing" - supportedTags={profileFilters} - query={query} - onSearch={handleSearch} - maxQueryLength={MAX_QUERY_LENGTH} - /> + {profilingUsingTransactions ? ( + <SearchBar + searchSource="profile_summary" + organization={organization} + projectIds={eventView.project} + query={query} + onSearch={handleSearch} + maxQueryLength={MAX_QUERY_LENGTH} + /> + ) : ( + <SmartSearchBar + organization={organization} + hasRecentSearches + searchSource="profile_landing" + supportedTags={profileFilters} + query={query} + onSearch={handleSearch} + maxQueryLength={MAX_QUERY_LENGTH} + /> + )} </ActionBar> {shouldShowProfilingOnboardingPanel ? ( <ProfilingOnboardingPanel> diff --git a/static/app/views/profiling/profileSummary/content.tsx b/static/app/views/profiling/profileSummary/content.tsx index 0a00fe2c1c730e..d66e92a97aca73 100644 --- a/static/app/views/profiling/profileSummary/content.tsx +++ b/static/app/views/profiling/profileSummary/content.tsx @@ -1,4 +1,4 @@ -import {useCallback, useMemo} from 'react'; +import {Fragment, useCallback, useMemo} from 'react'; import {browserHistory} from 'react-router'; import styled from '@emotion/styled'; import {Location} from 'history'; @@ -67,37 +67,39 @@ function ProfileSummaryContent(props: ProfileSummaryContentProps) { ); return ( - <Layout.Main fullWidth> - <ProfileCharts query={props.query} hideCount /> - <TableHeader> - <CompactSelect - triggerProps={{prefix: t('Filter'), size: 'xs'}} - value={sort.order === 'asc' ? sort.key : `-${sort.key}`} - options={FILTER_OPTIONS} - onChange={opt => handleFilterChange(opt.value)} + <Fragment> + <Layout.Main fullWidth> + <ProfileCharts query={props.query} hideCount /> + <TableHeader> + <CompactSelect + triggerProps={{prefix: t('Filter'), size: 'xs'}} + value={sort.order === 'asc' ? sort.key : `-${sort.key}`} + options={FILTER_OPTIONS} + onChange={opt => handleFilterChange(opt.value)} + /> + <StyledPagination + pageLinks={ + profiles.status === 'success' + ? profiles.data?.[2]?.getResponseHeader('Link') ?? null + : null + } + size="xs" + /> + </TableHeader> + <ProfileEventsTable + columns={fields} + data={profiles.status === 'success' ? profiles.data[0] : null} + error={profiles.status === 'error' ? t('Unable to load profiles') : null} + isLoading={profiles.status === 'loading'} + sort={sort} /> - <StyledPagination - pageLinks={ - profiles.status === 'success' - ? profiles.data?.[2]?.getResponseHeader('Link') ?? null - : null - } - size="xs" + <SuspectFunctionsTable + project={props.project} + transaction={props.transaction} + analyticsPageSource="profiling_transaction" /> - </TableHeader> - <ProfileEventsTable - columns={fields} - data={profiles.status === 'success' ? profiles.data[0] : null} - error={profiles.status === 'error' ? t('Unable to load profiles') : null} - isLoading={profiles.status === 'loading'} - sort={sort} - /> - <SuspectFunctionsTable - project={props.project} - transaction={props.transaction} - analyticsPageSource="profiling_transaction" - /> - </Layout.Main> + </Layout.Main> + </Fragment> ); } diff --git a/static/app/views/profiling/profileSummary/index.tsx b/static/app/views/profiling/profileSummary/index.tsx index 1a9c5390c17098..1acb0d3da8ec80 100644 --- a/static/app/views/profiling/profileSummary/index.tsx +++ b/static/app/views/profiling/profileSummary/index.tsx @@ -5,6 +5,7 @@ import {Location} from 'history'; import DatePageFilter from 'sentry/components/datePageFilter'; import EnvironmentPageFilter from 'sentry/components/environmentPageFilter'; +import SearchBar from 'sentry/components/events/searchBar'; import IdBadge from 'sentry/components/idBadge'; import * as Layout from 'sentry/components/layouts/thirds'; import PageFilterBar from 'sentry/components/organizations/pageFilterBar'; @@ -21,6 +22,7 @@ import {space} from 'sentry/styles/space'; import {PageFilters, Project} from 'sentry/types'; import {defined} from 'sentry/utils'; import trackAdvancedAnalyticsEvent from 'sentry/utils/analytics/trackAdvancedAnalyticsEvent'; +import EventView from 'sentry/utils/discover/eventView'; import {isAggregateField} from 'sentry/utils/discover/fields'; import {useCurrentProjectFromRouteParam} from 'sentry/utils/profiling/hooks/useCurrentProjectFromRouteParam'; import {useProfileFilters} from 'sentry/utils/profiling/hooks/useProfileFilters'; @@ -43,6 +45,10 @@ function ProfileSummaryPage(props: ProfileSummaryPageProps) { const organization = useOrganization(); const project = useCurrentProjectFromRouteParam(); + const profilingUsingTransactions = organization.features.includes( + 'profiling-using-transactions' + ); + useEffect(() => { trackAdvancedAnalyticsEvent('profiling_views.profile_summary', { organization, @@ -94,6 +100,7 @@ function ProfileSummaryPage(props: ProfileSummaryPageProps) { const profileFilters = useProfileFilters({ query: filtersQuery, selection: props.selection, + disabled: profilingUsingTransactions, }); const handleSearch: SmartSearchBarProps['onSearch'] = useCallback( @@ -129,6 +136,22 @@ function ProfileSummaryPage(props: ProfileSummaryPageProps) { ]; }, [props.location.query, project?.slug, transaction]); + const eventView = useMemo(() => { + const _eventView = EventView.fromNewQueryWithLocation( + { + id: undefined, + version: 2, + name: transaction || '', + fields: [], + query, + projects: project ? [parseInt(project.id, 10)] : [], + }, + props.location + ); + _eventView.additionalConditions.setFilterValues('has', ['profile.id']); + return _eventView; + }, [props.location, project, query, transaction]); + return ( <SentryDocumentTitle title={t('Profiling \u2014 Profile Summary')} @@ -162,21 +185,32 @@ function ProfileSummaryPage(props: ProfileSummaryPageProps) { </Layout.HeaderContent> </Layout.Header> <Layout.Body> - <Layout.Main fullWidth> + <Layout.Main fullWidth={!profilingUsingTransactions}> <ActionBar> <PageFilterBar condensed> <EnvironmentPageFilter /> <DatePageFilter alignDropdown="left" /> </PageFilterBar> - <SmartSearchBar - organization={organization} - hasRecentSearches - searchSource="profile_summary" - supportedTags={profileFilters} - query={rawQuery} - onSearch={handleSearch} - maxQueryLength={MAX_QUERY_LENGTH} - /> + {profilingUsingTransactions ? ( + <SearchBar + searchSource="profile_summary" + organization={organization} + projectIds={eventView.project} + query={rawQuery} + onSearch={handleSearch} + maxQueryLength={MAX_QUERY_LENGTH} + /> + ) : ( + <SmartSearchBar + organization={organization} + hasRecentSearches + searchSource="profile_summary" + supportedTags={profileFilters} + query={rawQuery} + onSearch={handleSearch} + maxQueryLength={MAX_QUERY_LENGTH} + /> + )} </ActionBar> <ProfileSummaryContent location={props.location}
dd7c37d4dae638611e20b76d5bc0653b1803e591
2024-04-10 12:50:23
Ogi
fix(dashboards): disable metric chart zoom (#68595)
false
disable metric chart zoom (#68595)
fix
diff --git a/static/app/views/dashboards/metrics/chart.tsx b/static/app/views/dashboards/metrics/chart.tsx index 04a62542da703a..7db8cc599037aa 100644 --- a/static/app/views/dashboards/metrics/chart.tsx +++ b/static/app/views/dashboards/metrics/chart.tsx @@ -49,7 +49,7 @@ export function MetricChartContainer({ displayType={displayType} group={DASHBOARD_CHART_GROUP} height={chartHeight} - enableZoom + enableZoom={false} /> </TransitionChart> </MetricWidgetChartWrapper>
e0704b75a6d7fe9547f6bdece1fcc24a730a9552
2024-02-29 17:21:26
ArthurKnaus
fix(alerts): Interval rounding error (#66057)
false
Interval rounding error (#66057)
fix
diff --git a/static/app/views/alerts/rules/metric/create.spec.tsx b/static/app/views/alerts/rules/metric/create.spec.tsx index 3f1e2e7828fe7f..5dc18d063c94e1 100644 --- a/static/app/views/alerts/rules/metric/create.spec.tsx +++ b/static/app/views/alerts/rules/metric/create.spec.tsx @@ -67,7 +67,7 @@ describe('Incident Rules Create', function () { interval: '60m', project: [2], query: 'event.type:error', - statsPeriod: '9999m', + statsPeriod: '9998m', yAxis: 'count()', referrer: 'api.organization-event-stats', }, diff --git a/static/app/views/alerts/rules/metric/metricRulePresets.tsx b/static/app/views/alerts/rules/metric/metricRulePresets.tsx index e47bb66db5bd63..81c6ce17ed9465 100644 --- a/static/app/views/alerts/rules/metric/metricRulePresets.tsx +++ b/static/app/views/alerts/rules/metric/metricRulePresets.tsx @@ -57,9 +57,9 @@ export function makeDefaultCta({ start: timePeriod.start, end: timePeriod.end, utc: timePeriod.utc, - // 7 days are 9999m in alerts as of a rounding error in the `events-stats` endpoint + // 7 days are 9998m in alerts as of a rounding error in the `events-stats` endpoint // We need to round to 7d here to display it correctly in DDM - statsPeriod: timePeriod.period === '9999m' ? '7d' : timePeriod.period, + statsPeriod: timePeriod.period === '9998m' ? '7d' : timePeriod.period, project: projects .filter(({slug}) => rule.projects.includes(slug)) .map(project => project.id), diff --git a/static/app/views/alerts/rules/metric/triggers/chart/index.spec.tsx b/static/app/views/alerts/rules/metric/triggers/chart/index.spec.tsx index ddca3125f8f6bb..5789cf6f7a8d5d 100644 --- a/static/app/views/alerts/rules/metric/triggers/chart/index.spec.tsx +++ b/static/app/views/alerts/rules/metric/triggers/chart/index.spec.tsx @@ -65,7 +65,7 @@ describe('Incident Rules Create', () => { interval: '1m', project: [2], query: 'event.type:error', - statsPeriod: '9999m', + statsPeriod: '9998m', yAxis: 'count()', referrer: 'api.organization-event-stats', }, @@ -78,7 +78,7 @@ describe('Incident Rules Create', () => { query: { project: ['2'], query: 'event.type:error', - statsPeriod: '9999m', + statsPeriod: '9998m', environment: [], }, }) @@ -119,7 +119,7 @@ describe('Incident Rules Create', () => { interval: '1m', project: [2], query: 'event.type:error', - statsPeriod: '9999m', + statsPeriod: '9998m', yAxis: 'count()', referrer: 'api.organization-event-stats', }, @@ -132,7 +132,7 @@ describe('Incident Rules Create', () => { query: { project: ['2'], query: 'event.type:error', - statsPeriod: '9999m', + statsPeriod: '9998m', environment: [], }, }) diff --git a/static/app/views/alerts/rules/metric/types.tsx b/static/app/views/alerts/rules/metric/types.tsx index 0753f8e136a837..c471e81bd998ce 100644 --- a/static/app/views/alerts/rules/metric/types.tsx +++ b/static/app/views/alerts/rules/metric/types.tsx @@ -126,9 +126,9 @@ export enum TimePeriod { ONE_DAY = '1d', THREE_DAYS = '3d', // Seven days is actually 10080m but Snuba can only return up to 10000 entries, for this - // we approximate to 9999m which prevents rounding errors due to the minutes granularity + // we approximate to 9998m which prevents rounding errors due to the minutes granularity // limitations. - SEVEN_DAYS = '9999m', + SEVEN_DAYS = '9998m', FOURTEEN_DAYS = '14d', }
e591e1aa39627bf137b54990460c12dd38820a68
2021-11-24 04:00:24
Stephen Cefali
ref(dashboards): adds source prop dashboard widget (#30148)
false
adds source prop dashboard widget (#30148)
ref
diff --git a/static/app/components/modals/addDashboardWidgetModal.tsx b/static/app/components/modals/addDashboardWidgetModal.tsx index dcb33f9c574c7a..3fb7bb8fb86f54 100644 --- a/static/app/components/modals/addDashboardWidgetModal.tsx +++ b/static/app/components/modals/addDashboardWidgetModal.tsx @@ -38,6 +38,7 @@ import {DISPLAY_TYPE_CHOICES} from 'app/views/dashboardsV2/data'; import { DashboardDetails, DashboardListItem, + DashboardWidgetSource, DisplayType, MAX_WIDGETS, Widget, @@ -67,8 +68,7 @@ export type DashboardWidgetModalOptions = { defaultTableColumns?: readonly string[]; defaultTitle?: string; displayType?: DisplayType; - fromDiscover?: boolean; - fromLibrary?: boolean; + source: DashboardWidgetSource; start?: DateString; end?: DateString; statsPeriod?: RelativePeriod | string; @@ -111,7 +111,7 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { constructor(props: Props) { super(props); - const {widget, defaultWidgetQuery, defaultTitle, displayType, fromDiscover} = props; + const {widget, defaultWidgetQuery, defaultTitle, displayType} = props; if (!widget) { this.state = { title: defaultTitle ?? '', @@ -119,7 +119,7 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { interval: '5m', queries: [defaultWidgetQuery ? {...defaultWidgetQuery} : {...newQuery}], errors: undefined, - loading: !!fromDiscover, + loading: !!this.omitDashboardProp, dashboards: [], userHasModified: false, widgetType: WidgetType.DISCOVER, @@ -141,13 +141,24 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { } componentDidMount() { - const {fromDiscover} = this.props; - if (fromDiscover) { + if (this.omitDashboardProp) { this.fetchDashboards(); } this.handleDefaultFields(); } + get omitDashboardProp() { + // when opening from discover or issues page, the user selects the dashboard in the widget UI + return [ + DashboardWidgetSource.DISCOVERV2, + DashboardWidgetSource.ISSUE_DETAILS, + ].includes(this.props.source); + } + + get fromLibrary() { + return this.props.source === DashboardWidgetSource.LIBRARY; + } + handleSubmit = async (event: React.FormEvent) => { event.preventDefault(); @@ -158,8 +169,7 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { onAddWidget, onUpdateWidget, widget: previousWidget, - fromDiscover, - fromLibrary, + source, } = this.props; this.setState({loading: true}); let errors: FlatValidationError = {}; @@ -194,7 +204,7 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { organization, }); } - if (!fromDiscover && !fromLibrary) { + if (source === DashboardWidgetSource.DASHBOARDS) { closeModal(); } } catch (err) { @@ -202,16 +212,19 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { this.setState({errors}); } finally { this.setState({loading: false}); - if (fromDiscover) { - this.handleSubmitFromDiscover(errors, widgetData); + if (this.omitDashboardProp) { + this.handleSubmitFromSelectedDashboard(errors, widgetData); } - if (fromLibrary) { + if (this.fromLibrary) { this.handleSubmitFromLibrary(errors, widgetData); } } }; - handleSubmitFromDiscover = async (errors: FlatValidationError, widgetData: Widget) => { + handleSubmitFromSelectedDashboard = async ( + errors: FlatValidationError, + widgetData: Widget + ) => { const {closeModal, organization} = this.props; const {selectedDashboard, dashboards} = this.state; // Validate that a dashboard was selected since api call to /dashboards/widgets/ does not check for dashboard @@ -315,13 +328,13 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { }; handleFieldChange = (field: string) => (value: string) => { - const {fromDiscover, organization} = this.props; + const {organization, source} = this.props; this.setState(prevState => { const newState = cloneDeep(prevState); set(newState, field, value); trackAdvancedAnalyticsEvent('dashboards_views.add_widget_modal.change', { - from: fromDiscover ? 'discoverv2' : 'dashboards', + from: source, field, value, widgetType: 'discover', @@ -471,8 +484,6 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { organization, selection, tags, - fromDiscover, - fromLibrary, widget: previousWidget, start, end, @@ -499,14 +510,13 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { }); const isUpdatingWidget = typeof onUpdateWidget === 'function' && !!previousWidget; - return ( <React.Fragment> <Header closeButton> <h4> - {fromDiscover + {this.omitDashboardProp ? t('Add Widget to Dashboard') - : fromLibrary + : this.fromLibrary ? t('Add Custom Widget') : isUpdatingWidget ? t('Edit Widget') @@ -514,7 +524,7 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { </h4> </Header> <Body> - {fromDiscover && this.renderDashboardSelector()} + {this.omitDashboardProp && this.renderDashboardSelector()} <DoubleFieldWrapper> <StyledField data-test-id="widget-name" @@ -604,7 +614,7 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { {t('Read the docs')} </Button> <ButtonBar gap={1}> - {fromLibrary && dashboard && onAddLibraryWidget ? ( + {this.fromLibrary && dashboard && onAddLibraryWidget ? ( <Button data-test-id="back-to-library" type="button" @@ -629,7 +639,7 @@ class AddDashboardWidgetModal extends React.Component<Props, State> { disabled={state.loading} busy={state.loading} > - {fromLibrary + {this.fromLibrary ? t('Confirm') : isUpdatingWidget ? t('Update Widget') diff --git a/static/app/components/modals/dashboardWidgetLibraryModal/index.tsx b/static/app/components/modals/dashboardWidgetLibraryModal/index.tsx index 923a9055e26732..b43e48a18b4565 100644 --- a/static/app/components/modals/dashboardWidgetLibraryModal/index.tsx +++ b/static/app/components/modals/dashboardWidgetLibraryModal/index.tsx @@ -8,7 +8,11 @@ import Tag from 'app/components/tagDeprecated'; import {t} from 'app/locale'; import space from 'app/styles/space'; import {Organization} from 'app/types'; -import {DashboardDetails, Widget} from 'app/views/dashboardsV2/types'; +import { + DashboardDetails, + DashboardWidgetSource, + Widget, +} from 'app/views/dashboardsV2/types'; import {WidgetTemplate} from 'app/views/dashboardsV2/widgetLibrary/data'; import Button from '../../button'; @@ -67,7 +71,7 @@ function DashboardWidgetLibraryModal({ dashboard, selectedWidgets, widget: customWidget, - fromLibrary: true, + source: DashboardWidgetSource.LIBRARY, onAddLibraryWidget: onAddWidget, }); }} diff --git a/static/app/views/dashboardsV2/dashboard.tsx b/static/app/views/dashboardsV2/dashboard.tsx index f209a979dae499..2d218dfacf5c1e 100644 --- a/static/app/views/dashboardsV2/dashboard.tsx +++ b/static/app/views/dashboardsV2/dashboard.tsx @@ -22,7 +22,13 @@ import withGlobalSelection from 'app/utils/withGlobalSelection'; import {DataSet} from './widget/utils'; import AddWidget, {ADD_WIDGET_BUTTON_DRAG_ID} from './addWidget'; import SortableWidget from './sortableWidget'; -import {DashboardDetails, MAX_WIDGETS, Widget, WidgetType} from './types'; +import { + DashboardDetails, + DashboardWidgetSource, + MAX_WIDGETS, + Widget, + WidgetType, +} from './types'; type Props = { api: Client; @@ -93,6 +99,7 @@ class Dashboard extends Component<Props> { dashboard, selection, onAddWidget: this.handleAddComplete, + source: DashboardWidgetSource.DASHBOARDS, }); }; @@ -182,6 +189,7 @@ class Dashboard extends Component<Props> { openAddDashboardWidgetModal({ ...modalProps, dashboard, + source: DashboardWidgetSource.DASHBOARDS, }); } }; diff --git a/static/app/views/dashboardsV2/types.tsx b/static/app/views/dashboardsV2/types.tsx index 5fd25986673fa5..247edaa419742a 100644 --- a/static/app/views/dashboardsV2/types.tsx +++ b/static/app/views/dashboardsV2/types.tsx @@ -66,3 +66,11 @@ export enum DashboardState { CREATE = 'create', PENDING_DELETE = 'pending_delete', } + +// where we launch the dashboard widget from +export enum DashboardWidgetSource { + DISCOVERV2 = 'discoverv2', + DASHBOARDS = 'dashboards', + LIBRARY = 'library', + ISSUE_DETAILS = 'issueDetail', +} diff --git a/static/app/views/eventsV2/queryList.tsx b/static/app/views/eventsV2/queryList.tsx index e7b6657f01f844..9e6af6a6ecb9a2 100644 --- a/static/app/views/eventsV2/queryList.tsx +++ b/static/app/views/eventsV2/queryList.tsx @@ -26,7 +26,7 @@ import {DisplayModes} from 'app/utils/discover/types'; import parseLinkHeader from 'app/utils/parseLinkHeader'; import {decodeList} from 'app/utils/queryString'; import withApi from 'app/utils/withApi'; -import {WidgetQuery} from 'app/views/dashboardsV2/types'; +import {DashboardWidgetSource, WidgetQuery} from 'app/views/dashboardsV2/types'; import { displayModeToDisplayType, @@ -122,7 +122,7 @@ class QueryList extends React.Component<Props> { start: eventView.start, end: eventView.end, statsPeriod: eventView.statsPeriod, - fromDiscover: true, + source: DashboardWidgetSource.DISCOVERV2, defaultWidgetQuery, defaultTableColumns: eventView.fields.map(({field}) => field), defaultTitle: diff --git a/static/app/views/eventsV2/savedQuery/index.tsx b/static/app/views/eventsV2/savedQuery/index.tsx index 3edc5c696b5ad6..7c9bbea2acc5be 100644 --- a/static/app/views/eventsV2/savedQuery/index.tsx +++ b/static/app/views/eventsV2/savedQuery/index.tsx @@ -26,7 +26,7 @@ import {DisplayModes} from 'app/utils/discover/types'; import {getDiscoverLandingUrl} from 'app/utils/discover/urls'; import withApi from 'app/utils/withApi'; import withProjects from 'app/utils/withProjects'; -import {WidgetQuery} from 'app/views/dashboardsV2/types'; +import {DashboardWidgetSource, WidgetQuery} from 'app/views/dashboardsV2/types'; import InputControl from 'app/views/settings/components/forms/controls/input'; import { @@ -246,7 +246,7 @@ class SavedQueryButtonGroup extends React.PureComponent<Props, State> { openAddDashboardWidgetModal({ organization, - fromDiscover: true, + source: DashboardWidgetSource.DISCOVERV2, defaultWidgetQuery, defaultTableColumns: eventView.fields.map(({field}) => field), defaultTitle: diff --git a/tests/js/spec/components/modals/addDashboardWidgetModal.spec.jsx b/tests/js/spec/components/modals/addDashboardWidgetModal.spec.jsx index b87ac4fb477f61..3e99ddc887a6ff 100644 --- a/tests/js/spec/components/modals/addDashboardWidgetModal.spec.jsx +++ b/tests/js/spec/components/modals/addDashboardWidgetModal.spec.jsx @@ -22,8 +22,7 @@ function mountModal({ onUpdateWidget, widget, dashboard, - fromDiscover, - fromLibrary, + source, defaultWidgetQuery, displayType, defaultTableColumns, @@ -41,8 +40,7 @@ function mountModal({ widget={widget} dashboard={dashboard} closeModal={() => void 0} - fromDiscover={fromDiscover} - fromLibrary={fromLibrary} + source={source || types.DashboardWidgetSource.DASHBOARDS} defaultWidgetQuery={defaultWidgetQuery} displayType={displayType} defaultTableColumns={defaultTableColumns} @@ -137,7 +135,10 @@ describe('Modals -> AddDashboardWidgetModal', function () { }); it('redirects correctly when creating a new dashboard', async function () { - const wrapper = mountModal({initialData, fromDiscover: true}); + const wrapper = mountModal({ + initialData, + source: types.DashboardWidgetSource.DISCOVERV2, + }); await tick(); await wrapper.update(); selectDashboard(wrapper, {label: t('+ Create New Dashboard'), value: 'new'}); @@ -151,7 +152,10 @@ describe('Modals -> AddDashboardWidgetModal', function () { }); it('redirects correctly when choosing an existing dashboard', async function () { - const wrapper = mountModal({initialData, fromDiscover: true}); + const wrapper = mountModal({ + initialData, + source: types.DashboardWidgetSource.DISCOVERV2, + }); await tick(); await wrapper.update(); selectDashboard(wrapper, {label: t('Test Dashboard'), value: '1'}); @@ -166,7 +170,10 @@ describe('Modals -> AddDashboardWidgetModal', function () { it('disables dashboards with max widgets', async function () { types.MAX_WIDGETS = 1; - const wrapper = mountModal({initialData, fromDiscover: true}); + const wrapper = mountModal({ + initialData, + source: types.DashboardWidgetSource.DISCOVERV2, + }); await tick(); await wrapper.update(); openMenu(wrapper, {name: 'dashboard', control: true}); @@ -931,7 +938,7 @@ describe('Modals -> AddDashboardWidgetModal', function () { onAddWidget: () => undefined, onUpdateWidget: () => undefined, widget: undefined, - fromDiscover: true, + source: types.DashboardWidgetSource.DISCOVERV2, defaultWidgetQuery: { name: '', fields: ['count()', 'failure_count()', 'count_unique(user)'], @@ -954,7 +961,7 @@ describe('Modals -> AddDashboardWidgetModal', function () { initialData, onAddWidget: () => undefined, onUpdateWidget: () => undefined, - fromDiscover: true, + source: types.DashboardWidgetSource.DISCOVERV2, displayType: types.DisplayType.BAR, }); @@ -967,7 +974,7 @@ describe('Modals -> AddDashboardWidgetModal', function () { initialData, onAddWidget: () => undefined, onUpdateWidget: () => undefined, - fromDiscover: true, + source: types.DashboardWidgetSource.DISCOVERV2, displayType: types.DisplayType.TOP_N, defaultWidgetQuery: {fields: ['count_unique(user)'], orderby: '-count_unique_user'}, defaultTableColumns: ['title', 'count()'], @@ -991,7 +998,7 @@ describe('Modals -> AddDashboardWidgetModal', function () { initialData, dashboard, onAddLibraryWidget: onAddLibraryWidgetMock, - fromLibrary: true, + source: types.DashboardWidgetSource.LIBRARY, }); const input = wrapper.find('Input[name="title"] input'); @@ -1008,7 +1015,7 @@ describe('Modals -> AddDashboardWidgetModal', function () { initialData, dashboard, onAddLibraryWidget: onAddLibraryWidgetMock, - fromLibrary: true, + source: types.DashboardWidgetSource.LIBRARY, }); expect(wrapper.find('Button[data-test-id="back-to-library"]')).toHaveLength(1);
dabcc8ed45645d5750995158412c878315974f1f
2018-01-30 03:38:47
Matt Robenolt
fix(api): Replace str.format with unicode.format
false
Replace str.format with unicode.format
fix
diff --git a/src/sentry/api/endpoints/organization_details.py b/src/sentry/api/endpoints/organization_details.py index b07e27ccc2ec00..71e00050b292b2 100644 --- a/src/sentry/api/endpoints/organization_details.py +++ b/src/sentry/api/endpoints/organization_details.py @@ -164,7 +164,7 @@ def save(self): # check if ORG_OPTIONS changed if option_inst.has_changed('value'): old_val = option_inst.old_value('value') - changed_data[key] = 'from {} to {}'.format(old_val, option_inst.value) + changed_data[key] = u'from {} to {}'.format(old_val, option_inst.value) option_inst.save() if 'openMembership' in self.init_data: @@ -199,12 +199,12 @@ def save(self): if f is not 'flag_field': if org.has_changed(f): old_val = org.old_value(f) - changed_data[f] = 'from {} to {}'.format(old_val, v) + changed_data[f] = u'from {} to {}'.format(old_val, v) else: # check if flag fields changed for f, v in six.iteritems(org_tracked_field['flag_field']): if org.flag_has_changed(f): - changed_data[f] = 'to {}'.format(v) + changed_data[f] = u'to {}'.format(v) org.save()
de353d9d88c239dae4028c26d389770080c19aad
2019-10-29 01:32:04
Manu
feat(outcomes): Support 10s granularity (#15280)
false
Support 10s granularity (#15280)
feat
diff --git a/src/sentry/tsdb/snuba.py b/src/sentry/tsdb/snuba.py index d3ca46f8ecdbcb..1b153baef9476b 100644 --- a/src/sentry/tsdb/snuba.py +++ b/src/sentry/tsdb/snuba.py @@ -104,6 +104,41 @@ class SnubaTSDB(BaseTSDB): ), } + # The Outcomes dataset aggregates outcomes into chunks of an hour. So, for rollups less than an hour, we want to + # query the raw outcomes dataset, with a few different settings (defined in lower_rollup_query_settings). + lower_rollup_query_settings = { + TSDBModel.organization_total_received: SnubaModelQuerySettings( + snuba.Dataset.OutcomesRaw, "org_id", None, [["outcome", "!=", outcomes.Outcome.INVALID]] + ), + TSDBModel.organization_total_rejected: SnubaModelQuerySettings( + snuba.Dataset.OutcomesRaw, + "org_id", + None, + [["outcome", "=", outcomes.Outcome.RATE_LIMITED]], + ), + TSDBModel.organization_total_blacklisted: SnubaModelQuerySettings( + snuba.Dataset.OutcomesRaw, "org_id", None, [["outcome", "=", outcomes.Outcome.FILTERED]] + ), + TSDBModel.project_total_received: SnubaModelQuerySettings( + snuba.Dataset.OutcomesRaw, + "project_id", + None, + [["outcome", "!=", outcomes.Outcome.INVALID]], + ), + TSDBModel.project_total_rejected: SnubaModelQuerySettings( + snuba.Dataset.OutcomesRaw, + "project_id", + None, + [["outcome", "=", outcomes.Outcome.RATE_LIMITED]], + ), + TSDBModel.project_total_blacklisted: SnubaModelQuerySettings( + snuba.Dataset.OutcomesRaw, + "project_id", + None, + [["outcome", "=", outcomes.Outcome.FILTERED]], + ), + } + all_model_query_settings = dict( model_columns.items() + model_being_upgraded_query_settings.items() ) @@ -129,7 +164,11 @@ def get_data( `group_on_time`: whether to add a GROUP BY clause on the 'time' field. `group_on_model`: whether to add a GROUP BY clause on the primary model. """ - model_query_settings = self.all_model_query_settings.get(model) + # 10s is the only rollup under an hour that we support + if rollup and rollup == 10 and model in self.lower_rollup_query_settings.keys(): + model_query_settings = self.lower_rollup_query_settings.get(model) + else: + model_query_settings = self.all_model_query_settings.get(model) if model_query_settings is None: raise Exception(u"Unsupported TSDBModel: {}".format(model.name)) @@ -229,7 +268,11 @@ def trim(self, result, groups, keys): del result[rk] def get_range(self, model, keys, start, end, rollup=None, environment_ids=None): - model_query_settings = self.all_model_query_settings.get(model) + # 10s is the only rollup under an hour that we support + if rollup and rollup == 10 and model in self.lower_rollup_query_settings.keys(): + model_query_settings = self.lower_rollup_query_settings.get(model) + else: + model_query_settings = self.all_model_query_settings.get(model) assert model_query_settings is not None, u"Unsupported TSDBModel: {}".format(model.name) diff --git a/src/sentry/utils/snuba.py b/src/sentry/utils/snuba.py index 6cae5df4a24813..a687c2159a8fe8 100644 --- a/src/sentry/utils/snuba.py +++ b/src/sentry/utils/snuba.py @@ -63,6 +63,7 @@ class Dataset(Enum): Events = "events" Transactions = "transactions" Outcomes = "outcomes" + OutcomesRaw = "outcomes_raw" DATASETS = {Dataset.Events: SENTRY_SNUBA_MAP, Dataset.Transactions: TRANSACTIONS_SENTRY_SNUBA_MAP} @@ -605,7 +606,7 @@ def _prepare_query_params(query_params): if query_params.dataset in [Dataset.Events, Dataset.Transactions]: (organization_id, params_to_update) = get_query_params_to_update_for_projects(query_params) - elif query_params.dataset == Dataset.Outcomes: + elif query_params.dataset in [Dataset.Outcomes, Dataset.OutcomesRaw]: (organization_id, params_to_update) = get_query_params_to_update_for_organizations( query_params ) diff --git a/tests/sentry/tsdb/test_snuba.py b/tests/sentry/tsdb/test_snuba.py index 6c42d18b8d1ff3..4d5cb4d309d174 100644 --- a/tests/sentry/tsdb/test_snuba.py +++ b/tests/sentry/tsdb/test_snuba.py @@ -15,6 +15,14 @@ def floor_to_hour_epoch(value): return int(to_timestamp(value)) +def floor_to_10s_epoch(value): + seconds = value.second + floored_second = 10 * (seconds / 10) + + value = value.replace(second=floored_second, microsecond=0) + return int(to_timestamp(value)) + + class SnubaTSDBTest(OutcomesSnubaTest): def setUp(self): super(SnubaTSDBTest, self).setUp() @@ -29,12 +37,7 @@ def setUp(self): def test_organization_outcomes(self): other_organization = self.create_organization() - for tsdb_model, outcome in [ - (TSDBModel.organization_total_received, Outcome.ACCEPTED), - (TSDBModel.organization_total_rejected, Outcome.RATE_LIMITED), - (TSDBModel.organization_total_blacklisted, Outcome.FILTERED), - ]: - # Create all the outcomes we will be querying + for outcome in [Outcome.ACCEPTED, Outcome.RATE_LIMITED, Outcome.FILTERED]: self.store_outcomes( self.organization.id, self.project.id, outcome.value, self.start_time, 3 ) @@ -50,34 +53,34 @@ def test_organization_outcomes(self): self.organization.id, self.project.id, outcome.value, self.day_before_start_time, 6 ) + for tsdb_model, granularity, floor_func, start_time_count, day_later_count in [ + (TSDBModel.organization_total_received, 3600, floor_to_hour_epoch, 3 * 3, 4 * 3), + (TSDBModel.organization_total_rejected, 3600, floor_to_hour_epoch, 3, 4), + (TSDBModel.organization_total_blacklisted, 3600, floor_to_hour_epoch, 3, 4), + (TSDBModel.organization_total_received, 10, floor_to_10s_epoch, 3 * 3, 4 * 3), + (TSDBModel.organization_total_rejected, 10, floor_to_10s_epoch, 3, 4), + (TSDBModel.organization_total_blacklisted, 10, floor_to_10s_epoch, 3, 4), + ]: # Query SnubaTSDB response = self.db.get_range( - tsdb_model, [self.organization.id], self.start_time, self.now, 3600, None + tsdb_model, [self.organization.id], self.start_time, self.now, granularity, None ) # Assert that the response has values set for the times we expect, and nothing more assert self.organization.id in response.keys() response_dict = {k: v for (k, v) in response[self.organization.id]} - assert response_dict[floor_to_hour_epoch(self.start_time)] == 3 - assert response_dict[floor_to_hour_epoch(self.one_day_later)] == 4 + assert response_dict[floor_func(self.start_time)] == start_time_count + assert response_dict[floor_func(self.one_day_later)] == day_later_count for time, count in response[self.organization.id]: - if time not in [ - floor_to_hour_epoch(self.start_time), - floor_to_hour_epoch(self.one_day_later), - ]: + if time not in [floor_func(self.start_time), floor_func(self.one_day_later)]: assert count == 0 def test_project_outcomes(self): other_project = self.create_project(organization=self.organization) - for tsdb_model, outcome in [ - (TSDBModel.project_total_received, Outcome.ACCEPTED), - (TSDBModel.project_total_rejected, Outcome.RATE_LIMITED), - (TSDBModel.project_total_blacklisted, Outcome.FILTERED), - ]: - # Create all the outcomes we will be querying + for outcome in [Outcome.ACCEPTED, Outcome.RATE_LIMITED, Outcome.FILTERED]: self.store_outcomes( self.organization.id, self.project.id, outcome.value, self.start_time, 3 ) @@ -93,21 +96,25 @@ def test_project_outcomes(self): self.organization.id, self.project.id, outcome.value, self.day_before_start_time, 6 ) - # Query SnubaTSDB + for tsdb_model, granularity, floor_func, start_time_count, day_later_count in [ + (TSDBModel.project_total_received, 3600, floor_to_hour_epoch, 3 * 3, 4 * 3), + (TSDBModel.project_total_rejected, 3600, floor_to_hour_epoch, 3, 4), + (TSDBModel.project_total_blacklisted, 3600, floor_to_hour_epoch, 3, 4), + (TSDBModel.project_total_received, 10, floor_to_10s_epoch, 3 * 3, 4 * 3), + (TSDBModel.project_total_rejected, 10, floor_to_10s_epoch, 3, 4), + (TSDBModel.project_total_blacklisted, 10, floor_to_10s_epoch, 3, 4), + ]: response = self.db.get_range( - tsdb_model, [self.project.id], self.start_time, self.now, 3600, None + tsdb_model, [self.project.id], self.start_time, self.now, granularity, None ) # Assert that the response has values set for the times we expect, and nothing more assert self.project.id in response.keys() response_dict = {k: v for (k, v) in response[self.project.id]} - assert response_dict[floor_to_hour_epoch(self.start_time)] == 3 - assert response_dict[floor_to_hour_epoch(self.one_day_later)] == 4 + assert response_dict[floor_func(self.start_time)] == start_time_count + assert response_dict[floor_func(self.one_day_later)] == day_later_count for time, count in response[self.project.id]: - if time not in [ - floor_to_hour_epoch(self.start_time), - floor_to_hour_epoch(self.one_day_later), - ]: + if time not in [floor_func(self.start_time), floor_func(self.one_day_later)]: assert count == 0
689a09790f37454199873857cbf71f93f7dc3902
2020-03-17 20:49:07
Matej Minar
feat(ui): Add Artifacts tab to releasesV2 (#17722)
false
Add Artifacts tab to releasesV2 (#17722)
feat
diff --git a/src/sentry/static/sentry/app/views/releases/detail/releaseArtifacts.jsx b/src/sentry/static/sentry/app/views/releases/detail/releaseArtifacts.jsx index ce0e457d4d2a85..48c6ce9b96dcdc 100644 --- a/src/sentry/static/sentry/app/views/releases/detail/releaseArtifacts.jsx +++ b/src/sentry/static/sentry/app/views/releases/detail/releaseArtifacts.jsx @@ -1,7 +1,7 @@ -import {Flex} from 'reflexbox'; import PropTypes from 'prop-types'; import React from 'react'; import omit from 'lodash/omit'; +import styled from '@emotion/styled'; import {Panel, PanelHeader, PanelBody, PanelItem} from 'app/components/panels'; import {URL_PARAM} from 'app/constants/globalSelectionHeader'; @@ -21,11 +21,13 @@ import SentryTypes from 'app/sentryTypes'; import Tooltip from 'app/components/tooltip'; import withApi from 'app/utils/withApi'; import withOrganization from 'app/utils/withOrganization'; +import space from 'app/styles/space'; class ReleaseArtifacts extends React.Component { static propTypes = { organization: SentryTypes.Organization, api: PropTypes.object, + projectId: PropTypes.string, }; constructor() { @@ -43,11 +45,13 @@ class ReleaseArtifacts extends React.Component { } getFilesEndpoint() { - const {orgId, projectId, version} = this.props.params; - const encodedVersion = encodeURIComponent(version); + // ?? to temporarily support releases V1 and V2 + const {orgId, projectId, version, release} = this.props.params; + const encodedVersion = encodeURIComponent(version ?? release); + const project = projectId ?? this.props.projectId; - return projectId - ? `/projects/${orgId}/${projectId}/releases/${encodedVersion}/files/` + return project + ? `/projects/${orgId}/${project}/releases/${encodedVersion}/files/` : `/organizations/${orgId}/releases/${encodedVersion}/files/`; } @@ -120,28 +124,22 @@ class ReleaseArtifacts extends React.Component { <div> <Panel> <PanelHeader> - <Flex flex="7" pr={2}> - {t('Name')} - </Flex> - <Flex flex="2">{t('Distribution')}</Flex> - <Flex flex="3">{t('Size')}</Flex> + <NameColumn>{t('Name')}</NameColumn> + <DistributionColumn>{t('Distribution')}</DistributionColumn> + <SizeAndActionsColumn>{t('Size')}</SizeAndActionsColumn> </PanelHeader> <PanelBody> {this.state.fileList.map(file => ( <PanelItem key={file.id}> - <Flex - flex="7" - pr={2} - style={{wordWrap: 'break-word', wordBreak: 'break-all'}} - > + <NameColumn> <strong>{file.name || '(empty)'}</strong> - </Flex> - <Flex flex="2"> + </NameColumn> + <DistributionColumn> {file.dist || <span className="text-light">{t('None')}</span>} - </Flex> - <Flex flex="3" justifyContent="space-between"> + </DistributionColumn> + <SizeAndActionsColumn> <FileSize bytes={file.size} /> - <Flex alignItems="center"> + <AlignCenter> {access.has('project:write') ? ( <a href={ @@ -174,8 +172,8 @@ class ReleaseArtifacts extends React.Component { <span className="icon icon-trash" /> </LinkWithConfirmation> </div> - </Flex> - </Flex> + </AlignCenter> + </SizeAndActionsColumn> </PanelItem> ))} </PanelBody> @@ -186,5 +184,27 @@ class ReleaseArtifacts extends React.Component { } } +const NameColumn = styled('div')` + display: flex; + flex: 7; + padding-right: ${space(2)}; + word-wrap: break-word; + word-break: break-all; +`; +const DistributionColumn = styled('div')` + display: flex; + flex: 2; + padding-right: ${space(2)}; +`; +const SizeAndActionsColumn = styled('div')` + display: flex; + flex: 3; + justify-content: space-between; +`; +const AlignCenter = styled('div')` + display: flex; + align-items: center; +`; + export {ReleaseArtifacts}; export default withOrganization(withApi(ReleaseArtifacts)); diff --git a/src/sentry/static/sentry/app/views/releasesV2/detail/artifacts/index.tsx b/src/sentry/static/sentry/app/views/releasesV2/detail/artifacts/index.tsx index eea2c36c0c220f..28a8a924378f57 100644 --- a/src/sentry/static/sentry/app/views/releasesV2/detail/artifacts/index.tsx +++ b/src/sentry/static/sentry/app/views/releasesV2/detail/artifacts/index.tsx @@ -1,8 +1,55 @@ import React from 'react'; +import styled from '@emotion/styled'; +import {Params} from 'react-router/lib/Router'; +import {Location} from 'history'; -type Props = {}; +import {t} from 'app/locale'; +import {GlobalSelection} from 'app/types'; +import Alert from 'app/components/alert'; +import space from 'app/styles/space'; +import ReleaseArtifactsV1 from 'app/views/releases/detail/releaseArtifacts'; +import withGlobalSelection from 'app/utils/withGlobalSelection'; -// TODO(releasesV2): finish this component -const ReleaseArtifacts = ({}: Props) => <div>ReleaseArtifacts</div>; +import {ReleaseContext} from '..'; -export default ReleaseArtifacts; +type Props = { + params: Params; + location: Location; + selection: GlobalSelection; +}; + +const ReleaseArtifacts = ({params, location, selection}: Props) => ( + <ReleaseContext.Consumer> + {release => { + const project = release?.projects.find(p => p.id === selection.projects[0]); + // TODO(releasesV2): we will handle this later with forced project selector + if (!project) { + return null; + } + + return ( + <ContentBox> + <Alert type="warning"> + {t( + 'We are working on improving this experience, therefore Artifacts will be moving to Settings soon.' + )} + </Alert> + + <ReleaseArtifactsV1 + params={params} + location={location} + projectId={project.slug} + /> + </ContentBox> + ); + }} + </ReleaseContext.Consumer> +); + +const ContentBox = styled('div')` + padding: ${space(4)}; + flex: 1; + background-color: ${p => p.theme.white}; +`; + +export default withGlobalSelection(ReleaseArtifacts);
70ce3397d1f83afd7b43812c4fbbe0319855e2eb
2024-11-15 22:47:09
anthony sottile
ref: fix typing for sentry.testutils.fixtures (#80754)
false
fix typing for sentry.testutils.fixtures (#80754)
ref
diff --git a/pyproject.toml b/pyproject.toml index bc3ac9fa22f982..05ec3165170d78 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -319,7 +319,6 @@ module = [ "sentry.tasks.auth", "sentry.tasks.base", "sentry.testutils.cases", - "sentry.testutils.fixtures", "sentry.testutils.helpers.notifications", "sentry.utils.auth", "sentry.utils.committers", diff --git a/src/sentry/testutils/fixtures.py b/src/sentry/testutils/fixtures.py index 7e14d102ea0d6d..d16f50f126e2fd 100644 --- a/src/sentry/testutils/fixtures.py +++ b/src/sentry/testutils/fixtures.py @@ -1,6 +1,6 @@ from __future__ import annotations -from collections.abc import Mapping +from collections.abc import Iterable, Mapping from datetime import datetime, timedelta from typing import Any @@ -26,6 +26,7 @@ from sentry.monitors.models import Monitor, MonitorType, ScheduleType from sentry.organizations.services.organization import RpcOrganization from sentry.silo.base import SiloMode +from sentry.snuba.models import QuerySubscription from sentry.testutils.factories import Factories from sentry.testutils.helpers.datetime import before_now, iso_format from sentry.testutils.silo import assume_test_silo_mode @@ -44,7 +45,7 @@ from sentry.users.models.identity import Identity, IdentityProvider from sentry.users.models.user import User from sentry.users.services.user import RpcUser -from sentry.workflow_engine.models import DataSource, Detector, Workflow +from sentry.workflow_engine.models import DataSource, Detector, DetectorState, Workflow from sentry.workflow_engine.types import DetectorPriorityLevel @@ -183,36 +184,27 @@ def create_project_template(self, **kwargs) -> ProjectTemplate: def create_project_bookmark(self, project=None, *args, **kwargs): if project is None: project = self.project - return Factories.create_project_bookmark(project=project, *args, **kwargs) + return Factories.create_project_bookmark(project, *args, **kwargs) def create_project_key(self, project=None, *args, **kwargs): if project is None: project = self.project - return Factories.create_project_key(project=project, *args, **kwargs) + return Factories.create_project_key(project, *args, **kwargs) def create_project_rule(self, project=None, *args, **kwargs) -> Rule: if project is None: project = self.project return Factories.create_project_rule(project, *args, **kwargs) - def create_slack_project_rule( - self, project=None, integration_id=None, channel_id=None, channel_name=None, *args, **kwargs - ): + def create_slack_project_rule(self, project=None, *args, **kwargs): if project is None: project = self.project - return Factories.create_slack_project_rule( - project, - integration_id=integration_id, - channel_id=channel_id, - channel_name=channel_name, - *args, - **kwargs, - ) + return Factories.create_slack_project_rule(project, *args, **kwargs) - def create_release(self, project=None, user=None, *args, **kwargs): + def create_release(self, project=None, *args, **kwargs): if project is None: project = self.project - return Factories.create_release(project=project, user=user, *args, **kwargs) + return Factories.create_release(project, *args, **kwargs) def create_group_release(self, project: Project | None = None, *args, **kwargs) -> GroupRelease: if project is None: @@ -249,7 +241,7 @@ def create_code_mapping(self, project=None, repo=None, organization_integration= def create_repo(self, project=None, *args, **kwargs): if project is None: project = self.project - return Factories.create_repo(project=project, *args, **kwargs) + return Factories.create_repo(project, *args, **kwargs) def create_commit(self, *args, **kwargs): return Factories.create_commit(*args, **kwargs) @@ -277,7 +269,7 @@ def create_usersocialauth( user: User | None = None, provider: str | None = None, uid: str | None = None, - extra_data: Mapping[str, Any] | None = None, + extra_data: dict[str, Any] | None = None, ): if not user: user = self.user @@ -291,7 +283,7 @@ def store_event(self, *args, **kwargs) -> Event: def create_group(self, project=None, *args, **kwargs): if project is None: project = self.project - return Factories.create_group(project=project, *args, **kwargs) + return Factories.create_group(project, *args, **kwargs) def create_file(self, **kwargs): return Factories.create_file(**kwargs) @@ -302,12 +294,12 @@ def create_file_from_path(self, *args, **kwargs): def create_event_attachment(self, event=None, *args, **kwargs): if event is None: event = self.event - return Factories.create_event_attachment(event=event, *args, **kwargs) + return Factories.create_event_attachment(event, *args, **kwargs) - def create_dif_file(self, project=None, *args, **kwargs): + def create_dif_file(self, project: Project | None = None, *args, **kwargs): if project is None: project = self.project - return Factories.create_dif_file(project=project, *args, **kwargs) + return Factories.create_dif_file(project, *args, **kwargs) def create_dif_from_path(self, project=None, *args, **kwargs): if project is None: @@ -371,18 +363,16 @@ def create_integration_external_issue(self, *args, **kwargs): def create_integration_external_project(self, *args, **kwargs): return Factories.create_integration_external_project(*args, **kwargs) - def create_incident(self, organization=None, projects=None, subscription=None, *args, **kwargs): + def create_incident(self, organization=None, projects=None, *args, **kwargs): if not organization: organization = self.organization if projects is None: projects = [self.project] - return Factories.create_incident( - organization=organization, projects=projects, subscription=subscription, *args, **kwargs - ) + return Factories.create_incident(organization, projects, *args, **kwargs) - def create_incident_activity(self, incident, *args, **kwargs): - return Factories.create_incident_activity(incident=incident, *args, **kwargs) + def create_incident_activity(self, *args, **kwargs): + return Factories.create_incident_activity(*args, **kwargs) def create_incident_comment(self, incident, *args, **kwargs): return self.create_incident_activity( @@ -401,8 +391,8 @@ def create_alert_rule(self, organization=None, projects=None, *args, **kwargs) - def create_alert_rule_activation( self, - alert_rule=None, - query_subscriptions=None, + alert_rule: AlertRule | None = None, + query_subscriptions: Iterable[QuerySubscription] | None = None, project=None, monitor_type=AlertRuleMonitorTypeInt.ACTIVATED, activator=None, @@ -411,9 +401,7 @@ def create_alert_rule_activation( **kwargs, ): if not alert_rule: - alert_rule = self.create_alert_rule( - monitor_type=monitor_type, - ) + alert_rule = self.create_alert_rule(monitor_type=monitor_type) if not query_subscriptions: projects = [project] if project else [self.project] # subscribing an activated alert rule will create an activation @@ -427,9 +415,7 @@ def create_alert_rule_activation( created_activations = [] for sub in query_subscriptions: created_activations.append( - Factories.create_alert_rule_activation( - alert_rule=alert_rule, query_subscription=sub, *args, **kwargs - ) + Factories.create_alert_rule_activation(alert_rule, sub, *args, **kwargs) ) return created_activations @@ -622,9 +608,6 @@ def create_dashboard_widget(self, *args, **kwargs): def create_dashboard_widget_query(self, *args, **kwargs): return Factories.create_dashboard_widget_query(*args, **kwargs) - def create_workflow_action(self, *args, **kwargs) -> Workflow: - return Factories.create_workflow_action(*args, **kwargs) - def create_workflow(self, *args, **kwargs) -> Workflow: return Factories.create_workflow(*args, **kwargs) @@ -654,7 +637,7 @@ def create_data_condition( def create_detector(self, *args, **kwargs) -> Detector: return Factories.create_detector(*args, **kwargs) - def create_detector_state(self, *args, **kwargs) -> Detector: + def create_detector_state(self, *args, **kwargs) -> DetectorState: return Factories.create_detector_state(*args, **kwargs) def create_data_source_detector(self, *args, **kwargs):
cbfd2783371c113c1a6be79c4381c5e98ee8ec75
2023-10-27 03:28:44
Evan Purkhiser
fix(ui): Use inline CommitLink in resolutionBox (#58899)
false
Use inline CommitLink in resolutionBox (#58899)
fix
diff --git a/static/app/components/resolutionBox.tsx b/static/app/components/resolutionBox.tsx index 756a0d3b576b4c..f8ebd955dc6442 100644 --- a/static/app/components/resolutionBox.tsx +++ b/static/app/components/resolutionBox.tsx @@ -94,6 +94,8 @@ function renderReason( commit: ( <Fragment> <CommitLink + inline + showIcon={false} commitId={statusDetails.inCommit.id} repository={statusDetails.inCommit.repository as Repository} />
9f7a0160277a9f83db14cea62af9c09bb9e5ad93
2024-06-04 21:44:38
Cathy Teng
feat(slack): add slack-sdk (#71881)
false
add slack-sdk (#71881)
feat
diff --git a/requirements-base.txt b/requirements-base.txt index 594d59e5b890aa..1a69e063ef8b81 100644 --- a/requirements-base.txt +++ b/requirements-base.txt @@ -67,6 +67,7 @@ sentry-ophio==0.2.7 sentry-redis-tools>=0.1.7 sentry-relay>=0.8.67 sentry-sdk>=2.2.1 +slack-sdk>=3.27.2 snuba-sdk>=2.0.33 simplejson>=3.17.6 sqlparse>=0.4.4 diff --git a/requirements-dev-frozen.txt b/requirements-dev-frozen.txt index 1985330b769161..d52f04fe48881d 100644 --- a/requirements-dev-frozen.txt +++ b/requirements-dev-frozen.txt @@ -189,6 +189,7 @@ sentry-sdk==2.2.1 sentry-usage-accountant==0.0.10 simplejson==3.17.6 six==1.16.0 +slack-sdk==3.27.2 sniffio==1.2.0 snuba-sdk==2.0.34 sortedcontainers==2.4.0 diff --git a/requirements-frozen.txt b/requirements-frozen.txt index 9ba04893365fa5..bd621848f17765 100644 --- a/requirements-frozen.txt +++ b/requirements-frozen.txt @@ -129,6 +129,7 @@ sentry-sdk==2.2.1 sentry-usage-accountant==0.0.10 simplejson==3.17.6 six==1.16.0 +slack-sdk==3.27.2 sniffio==1.3.0 snuba-sdk==2.0.34 soupsieve==2.3.2.post1
9ea399f2d5def91eb7fd258ef8721a17ab9069c7
2023-04-26 00:16:57
Alberto Leal
fix(hybrid-cloud): Fix org mapping deletes (#47908)
false
Fix org mapping deletes (#47908)
fix
diff --git a/src/sentry/models/organization.py b/src/sentry/models/organization.py index e0c3b6b106058a..c1be76ce9b37b8 100644 --- a/src/sentry/models/organization.py +++ b/src/sentry/models/organization.py @@ -8,7 +8,6 @@ from django.conf import settings from django.db import IntegrityError, models, router, transaction from django.db.models import QuerySet -from django.db.models.signals import post_delete from django.urls import NoReverseMatch, reverse from django.utils import timezone from django.utils.functional import cached_property @@ -676,13 +675,6 @@ def get_teams_with_org_roles(self, roles: Optional[Collection[str]]) -> QuerySet return Team.objects.filter(organization=self).exclude(org_role=None) - # TODO(hybrid-cloud): Replace with Region tombstone when it's implemented - @classmethod - def remove_organization_mapping(cls, instance, **kwargs): - from sentry.services.hybrid_cloud.organization_mapping import organization_mapping_service - - organization_mapping_service.delete(organization_id=instance.id) - def organization_absolute_url( has_customer_domain: bool, @@ -716,11 +708,3 @@ def organization_absolute_url( if fragment: parts.append(fragment) return "".join(parts) - - -post_delete.connect( - Organization.remove_organization_mapping, - dispatch_uid="sentry.remove_organization_mapping", - sender=Organization, - weak=False, -) diff --git a/src/sentry/receivers/outbox/region.py b/src/sentry/receivers/outbox/region.py index 3b223ae8a8408d..1dd8e81756414b 100644 --- a/src/sentry/receivers/outbox/region.py +++ b/src/sentry/receivers/outbox/region.py @@ -79,6 +79,7 @@ def process_team_updates( @receiver(process_region_outbox, sender=OutboxCategory.ORGANIZATION_UPDATE) def process_organization_updates(object_identifier: int, **kwds: Any): if (org := maybe_process_tombstone(Organization, object_identifier)) is None: + organization_mapping_service.delete(organization_id=object_identifier) return update = update_organization_mapping_from_instance(org) diff --git a/tests/sentry/deletions/test_organization.py b/tests/sentry/deletions/test_organization.py index 40840a3dde7744..fd892b274e17d4 100644 --- a/tests/sentry/deletions/test_organization.py +++ b/tests/sentry/deletions/test_organization.py @@ -27,6 +27,7 @@ from sentry.snuba.models import SnubaQuery from sentry.tasks.deletion.scheduled import run_deletion from sentry.testutils import TransactionTestCase +from sentry.testutils.outbox import outbox_runner from sentry.testutils.silo import region_silo_test @@ -94,7 +95,7 @@ def test_simple(self): deletion = ScheduledDeletion.schedule(org, days=0) deletion.update(in_progress=True) - with self.tasks(): + with self.tasks(), outbox_runner(): run_deletion(deletion.id) assert Organization.objects.filter(id=org2.id).exists()
0628cec44cc06f9f8553d951858173c6ea8d15e6
2021-09-08 18:31:47
Armen Zambrano G
fix(ci): Remove automation to initialize Docker on Mac (#28352)
false
Remove automation to initialize Docker on Mac (#28352)
fix
diff --git a/.github/workflows/development-environment.yml b/.github/workflows/development-environment.yml index 00fb6bd36fdb31..6723bfa46ccf5b 100644 --- a/.github/workflows/development-environment.yml +++ b/.github/workflows/development-environment.yml @@ -3,6 +3,7 @@ on: pull_request: paths: - 'Makefile' + - '.github/actions/*' - '.github/workflows/development-environment.yml' - '.envrc' - 'Brewfile' @@ -17,7 +18,6 @@ jobs: timeout-minutes: 40 strategy: matrix: - # macosx-11.0 is Big Sur, however, it takes long for jobs to get started # Using Ubuntu 18 until I figure out this error: # -> ImportError: libffi.so.6: cannot open shared object file: No such file or directory os: [macos-11.0, ubuntu-18.04] @@ -31,30 +31,28 @@ jobs: - name: Checkout sentry uses: actions/checkout@v2 + - name: Set variables for caches + id: info + run: | + echo "::set-output name=brew-cache-dir::$(brew --cache)" + echo "::set-output name=yarn-cache-dir::$(yarn cache dir)" + + - name: Cache (brew) + uses: actions/cache@v2 + with: + path: ${{ steps.info.outputs.brew-cache-dir }} + key: devenv-${{ runner.os }}-brew-${{ hashFiles('Brewfile') }} + restore-keys: devenv-${{ runner.os }}-brew + - name: Install prerequisites + # brew can be finicky but it does not always means that the rest of the job will fail # Xcode CLI & brew are already installed, thus, no need to call xcode-select install # Sometimes, brew needs to be updated before brew bundle would work - # After installing Docker (via homebrew) we need to make sure that it is properly initialized on Mac run: | - brew update && brew bundle -q - # This code is mentioned in our dev docs. Only remove if you adjust the docs as well - SENTRY_NO_VENV_CHECK=1 ./scripts/do.sh init-docker + brew update -q && brew bundle -q || exit 0 - # The next few steps are to set up the cache quickly - - name: Set environment variables & others - id: info - run: | - echo "::set-output name=python-version::$(SENTRY_NO_VENV_CHECK=1 ./scripts/do.sh get-pyenv-version)" - echo "::set-output name=pip-cache-dir::$(pip3 cache dir)" - echo "::set-output name=pip-version::$(pip -V | awk -F ' ' '{print $2}')" - echo "::set-output name=yarn-cache-dir::$(yarn cache dir)" - - # In a sense, we set up Python two times (once here and once via pyenv). Setting - # it up here is instant and it helps us to get the cache primed sooner - name: Setup Python - uses: actions/setup-python@v2 - with: - python-version: ${{ steps.info.outputs.python-version }} + uses: ./.github/actions/setup-python - name: Cache (pyenv) uses: actions/cache@v2 @@ -62,12 +60,6 @@ jobs: path: ~/.pyenv key: devenv-${{ matrix.os }}-pyenv-${{ hashFiles('.python-version') }} - - name: Cache (pip) - uses: actions/cache@v2 - with: - path: ${{ steps.info.outputs.pip-cache-dir }} - key: devenv-${{ matrix.os }}-py${{ steps.info.outputs.python-version }}-pip${{ steps.info.outputs.pip-version }}-${{ hashFiles('**/requirements.txt') }} - - name: Cache (yarn) uses: actions/cache@v1 # We are explicitly using v1 due to perf reasons with: @@ -83,9 +75,4 @@ jobs: eval "$(pyenv init --path)" python -m venv .venv source .venv/bin/activate - make bootstrap - - - name: Test direnv - run: | - brew install direnv - direnv allow + make develop init-config diff --git a/.github/workflows/files/com.docker.vmnetd.plist b/.github/workflows/files/com.docker.vmnetd.plist deleted file mode 100644 index 27bb3848d9b010..00000000000000 --- a/.github/workflows/files/com.docker.vmnetd.plist +++ /dev/null @@ -1,28 +0,0 @@ -<?xml version="1.0" encoding="UTF-8"?> -<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> -<plist version="1.0"> -<dict> - <key>Label</key> - <string>com.docker.vmnetd</string> - <key>Program</key> - <string>/Library/PrivilegedHelperTools/com.docker.vmnetd</string> - <key>ProgramArguments</key> - <array> - <string>/Library/PrivilegedHelperTools/com.docker.vmnetd</string> - </array> - <key>RunAtLoad</key> - <true/> - <key>Sockets</key> - <dict> - <key>Listener</key> - <dict> - <key>SockPathMode</key> - <integer>438</integer> - <key>SockPathName</key> - <string>/var/run/com.docker.vmnetd.sock</string> - </dict> - </dict> - <key>Version</key> - <string>59</string> -</dict> -</plist> diff --git a/.github/workflows/python-deps.yml b/.github/workflows/python-deps.yml index 654d9b8a3c7a28..b76895729b9e9e 100644 --- a/.github/workflows/python-deps.yml +++ b/.github/workflows/python-deps.yml @@ -2,6 +2,7 @@ name: python deps on: pull_request: paths: + - '.github/actions/*' - '.github/workflows/python-deps.yml' - 'requirements*' @@ -26,17 +27,8 @@ jobs: steps: - uses: actions/checkout@v2 - - name: Install prerequisites - # Sometimes, brew needs to be updated before brew bundle would work - run: | - brew update && brew bundle -q - - - name: Setup python - id: setup-python + - name: Setup Python uses: ./.github/actions/setup-python - with: - # XXX: We need to pass this python-deps-${{ matrix.os }}-py${{ matrix.python-version }}-${{ hashFiles('requirements-*.txt') }} - cache-files-hash: ${{ hashFiles('requirements-*.txt') }} - name: Install dependencies run: | diff --git a/scripts/lib.sh b/scripts/lib.sh index 232fbda4cb975f..0a17ab66241d45 100755 --- a/scripts/lib.sh +++ b/scripts/lib.sh @@ -79,53 +79,6 @@ sudo-askpass() { fi } -# After using homebrew to install docker, we need to do some magic to remove the need to interact with the GUI -# See: https://github.com/docker/for-mac/issues/2359#issuecomment-607154849 for why we need to do things below -init-docker() { - # Need to start docker if it was freshly installed or updated - # You will know that Docker is ready for devservices when the icon on the menu bar stops flashing - if query-mac && ! require docker && [ -d "/Applications/Docker.app" ]; then - echo "Making some changes to complete Docker initialization" - # allow the app to run without confirmation - xattr -d -r com.apple.quarantine /Applications/Docker.app - - # preemptively do docker.app's setup to avoid any gui prompts - # This path is not available for brand new MacBooks - sudo-askpass /bin/mkdir -p /Library/PrivilegedHelperTools - sudo-askpass /bin/chmod 754 /Library/PrivilegedHelperTools - sudo-askpass /bin/cp /Applications/Docker.app/Contents/Library/LaunchServices/com.docker.vmnetd /Library/PrivilegedHelperTools/ - sudo-askpass /bin/chmod 544 /Library/PrivilegedHelperTools/com.docker.vmnetd - - # This file used to be generated as part of brew's installation - if [ -f /Applications/Docker.app/Contents/Resources/com.docker.vmnetd.plist ]; then - sudo-askpass /bin/cp /Applications/Docker.app/Contents/Resources/com.docker.vmnetd.plist /Library/LaunchDaemons/ - else - sudo-askpass /bin/cp .github/workflows/files/com.docker.vmnetd.plist /Library/LaunchDaemons/ - fi - sudo-askpass /bin/chmod 644 /Library/LaunchDaemons/com.docker.vmnetd.plist - sudo-askpass /bin/launchctl load /Library/LaunchDaemons/com.docker.vmnetd.plist - fi - start-docker -} - -# This is mainly to be used by CI -# We need this for Mac since the executable docker won't work properly -# until the app is opened once -start-docker() { - if query-mac && ! docker system info &>/dev/null; then - echo "About to open Docker.app" - # At a later stage in the script, we're going to execute - # ensure_docker_server which waits for it to be ready - if ! open -g -a Docker.app; then - # If the step above fails, at least we can get some debugging information to determine why - sudo-askpass ls -l /Library/PrivilegedHelperTools/com.docker.vmnetd - ls -l /Library/LaunchDaemons/ - cat /Library/LaunchDaemons/com.docker.vmnetd.plist - ls -l /Applications/Docker.app - fi - fi -} - upgrade-pip() { pip install --upgrade "pip==21.1.2" "wheel==0.36.2" }
bf5e841bf72fd6a1d11bc46470ec6a57847d0d1d
2022-11-16 01:44:37
edwardgou-sentry
feat(mobile-exp): adds some more analytics to tag summary interactions (#41389)
false
adds some more analytics to tag summary interactions (#41389)
feat
diff --git a/static/app/components/group/sidebar.tsx b/static/app/components/group/sidebar.tsx index 07d8ed77fad55c..3ae3b6ef036003 100644 --- a/static/app/components/group/sidebar.tsx +++ b/static/app/components/group/sidebar.tsx @@ -330,6 +330,18 @@ class BaseGroupSidebar extends Component<Props, State> { organization={organization} projectId={projectId} group={group} + onTagClick={(title, value) => { + trackAdvancedAnalyticsEvent( + 'issue_group_details.tags.bar.clicked', + { + tag: title, + value: value.name, + platform: project.platform, + is_mobile: isMobilePlatform(project?.platform), + organization, + } + ); + }} /> ); }) diff --git a/static/app/components/group/tagDistributionMeter.tsx b/static/app/components/group/tagDistributionMeter.tsx index 8a6e96db1387ba..0f629a71457a2d 100644 --- a/static/app/components/group/tagDistributionMeter.tsx +++ b/static/app/components/group/tagDistributionMeter.tsx @@ -7,6 +7,7 @@ import {Group, Organization, TagWithTopValues} from 'sentry/types'; type Props = { group: Group; name: string; + onTagClick: React.ComponentProps<typeof TagDistributionMeter>['onTagClick']; organization: Organization; projectId: string; tag: string; @@ -25,7 +26,7 @@ class GroupTagDistributionMeter extends Component<Props> { } render() { - const {organization, group, tag, totalValues, topValues} = this.props; + const {organization, group, tag, totalValues, topValues, onTagClick} = this.props; const url = `/organizations/${organization.slug}/issues/${group.id}/tags/${tag}/?referrer=tag-distribution-meter`; const segments = topValues @@ -43,6 +44,7 @@ class GroupTagDistributionMeter extends Component<Props> { isLoading={false} hasError={false} segments={segments} + onTagClick={onTagClick} /> ); } diff --git a/static/app/components/group/tagFacets/tagFacetsBars.tsx b/static/app/components/group/tagFacets/tagFacetsBars.tsx index 2367bdb1197109..674ef7709df76b 100644 --- a/static/app/components/group/tagFacets/tagFacetsBars.tsx +++ b/static/app/components/group/tagFacets/tagFacetsBars.tsx @@ -8,7 +8,7 @@ import * as SidebarSection from 'sentry/components/sidebarSection'; import Tooltip from 'sentry/components/tooltip'; import {t} from 'sentry/locale'; import space from 'sentry/styles/space'; -import {TagWithTopValues} from 'sentry/types'; +import {Project, TagWithTopValues} from 'sentry/types'; import trackAdvancedAnalyticsEvent from 'sentry/utils/analytics/trackAdvancedAnalyticsEvent'; import {formatPercentage} from 'sentry/utils/formatters'; import {isMobilePlatform} from 'sentry/utils/platform'; @@ -129,7 +129,12 @@ export default function TagFacetsBars({ ); })} </StyledButtonBar> - <BreakdownBars data={points} maxItems={MAX_ITEMS} /> + <BreakdownBars + tag={state.selectedTag} + data={points} + maxItems={MAX_ITEMS} + project={project} + /> <Button size="xs" to={getTagUrl(organization.slug, groupId)} @@ -174,9 +179,12 @@ type Props = { */ data: Point[]; maxItems?: number; + project?: Project; + tag?: string; }; -function BreakdownBars({data, maxItems}: Props) { +function BreakdownBars({data, maxItems, project, tag}: Props) { + const organization = useOrganization(); const total = data.reduce((sum, point) => point.value + sum, 0); return ( <BreakdownGrid> @@ -198,6 +206,17 @@ function BreakdownBars({data, maxItems}: Props) { <Link to={point.url} aria-label={t('Add %s to the search query', point.label)} + onClick={() => { + if (tag && project) { + trackAdvancedAnalyticsEvent('issue_group_details.tags.bar.clicked', { + tag, + value: point.label, + platform: project.platform, + is_mobile: isMobilePlatform(project?.platform), + organization, + }); + } + }} > {bar} </Link> diff --git a/static/app/utils/analytics/issueAnalyticsEvents.tsx b/static/app/utils/analytics/issueAnalyticsEvents.tsx index 1cd17c7f668eca..82f598a3090bab 100644 --- a/static/app/utils/analytics/issueAnalyticsEvents.tsx +++ b/static/app/utils/analytics/issueAnalyticsEvents.tsx @@ -36,6 +36,12 @@ export type IssueEventParameters = { os?: string; platform?: string; }; + 'issue_group_details.tags.bar.clicked': { + is_mobile: boolean; + tag: string; + value: string; + platform?: string; + }; 'issue_group_details.tags.show_all_tags.clicked': { is_mobile: boolean; tag: string; @@ -47,6 +53,12 @@ export type IssueEventParameters = { tag: string; platform?: string; }; + 'issue_group_details.tags_distribution.bar.clicked': { + is_mobile: boolean; + tag: string; + value: string; + platform?: string; + }; 'issue_search.empty': { query: string; search_source: string; @@ -140,6 +152,9 @@ export const issueEventMap: Record<IssueEventKey, string | null> = { 'Issue Group Details: Tags show all clicked', 'issue_group_details.tags.switcher.clicked': 'Issue Group Details: Tags switcher clicked', + 'issue_group_details.tags.bar.clicked': 'Issue Group Details: Tags value bar clicked', + 'issue_group_details.tags_distribution.bar.clicked': + 'Issue Group Details: Tags distribution value bar clicked', // Performance Issue specific events here 'issue_details.performance.autogrouped_siblings_toggle':
62642ab98ab3d62d029bfaf75fcc9bcf8ece8125
2024-04-27 00:06:56
Scott Cooper
feat(ui): Keep stable references to lazy components (#69740)
false
Keep stable references to lazy components (#69740)
feat
diff --git a/static/app/components/events/rrwebIntegration.tsx b/static/app/components/events/rrwebIntegration.tsx index 7e9aacba1118d8..1709c5d225e6d4 100644 --- a/static/app/components/events/rrwebIntegration.tsx +++ b/static/app/components/events/rrwebIntegration.tsx @@ -1,3 +1,4 @@ +import {lazy} from 'react'; import styled from '@emotion/styled'; import {EventDataSection} from 'sentry/components/events/eventDataSection'; @@ -18,6 +19,8 @@ type Props = { projectSlug: Project['slug']; }; +const LazyReplayer = lazy(() => import('./rrwebReplayer')); + function EventRRWebIntegrationContent({orgId, projectSlug, event}: Props) { const { data: attachmentList, @@ -61,7 +64,7 @@ function EventRRWebIntegrationContent({orgId, projectSlug, event}: Props) { return ( <StyledReplayEventDataSection type="context-replay" title={t('Replay')}> <LazyLoad - component={() => import('./rrwebReplayer')} + LazyComponent={LazyReplayer} urls={attachmentList.map(createAttachmentUrl)} /> </StyledReplayEventDataSection> diff --git a/static/app/components/feedback/feedbackItem/replaySection.tsx b/static/app/components/feedback/feedbackItem/replaySection.tsx index 428cd8a26358ae..ae2a9c06af376d 100644 --- a/static/app/components/feedback/feedbackItem/replaySection.tsx +++ b/static/app/components/feedback/feedbackItem/replaySection.tsx @@ -1,4 +1,4 @@ -import {useCallback} from 'react'; +import {lazy} from 'react'; import LazyLoad from 'sentry/components/lazyLoad'; import type {Organization} from 'sentry/types/organization'; @@ -15,20 +15,18 @@ const CLIP_OFFSETS = { durationBeforeMs: 20_000, }; +const LazyReplayClipPreviewComponent = lazy( + () => import('sentry/components/events/eventReplay/replayClipPreview') +); +const LazyReplayPreviewComponent = lazy( + () => import('sentry/components/events/eventReplay/replayPreview') +); + export default function ReplaySection({eventTimestampMs, organization, replayId}: Props) { const hasUserFeedbackReplayClip = organization.features.includes( 'user-feedback-replay-clip' ); - const replayPreview = useCallback( - () => import('sentry/components/events/eventReplay/replayPreview'), - [] - ); - const replayClipPreview = useCallback( - () => import('sentry/components/events/eventReplay/replayClipPreview'), - [] - ); - const props = { analyticsContext: 'feedback', eventTimestampMs, @@ -45,8 +43,12 @@ export default function ReplaySection({eventTimestampMs, organization, replayId} }; return hasUserFeedbackReplayClip ? ( - <LazyLoad {...props} component={replayClipPreview} clipOffsets={CLIP_OFFSETS} /> + <LazyLoad + {...props} + LazyComponent={LazyReplayClipPreviewComponent} + clipOffsets={CLIP_OFFSETS} + /> ) : ( - <LazyLoad {...props} component={replayPreview} /> + <LazyLoad {...props} LazyComponent={LazyReplayPreviewComponent} /> ); } diff --git a/static/app/components/lazyLoad.tsx b/static/app/components/lazyLoad.tsx index 00e5552482614c..3ffc23fb8518b4 100644 --- a/static/app/components/lazyLoad.tsx +++ b/static/app/components/lazyLoad.tsx @@ -14,10 +14,16 @@ type PromisedImport<C> = Promise<{default: C}>; type ComponentType = React.ComponentType<any>; type Props<C extends ComponentType> = React.ComponentProps<C> & { + /** + * Wrap the component with lazy() before passing it to LazyLoad. + */ + LazyComponent?: React.LazyExoticComponent<C>; + /** * Accepts a function to trigger the import resolution of the component. + * @deprecated Use `LazyComponent` instead and keep lazy() calls out of the render path. */ - component: () => PromisedImport<C>; + component?: () => PromisedImport<C>; /** * Override the default fallback component. @@ -31,17 +37,24 @@ type Props<C extends ComponentType> = React.ComponentProps<C> & { * LazyLoad is used to dynamically load codesplit components via a `import` * call. This is primarily used in our routing tree. * - * <LazyLoad component={() => import('./myComponent')} someComponentProps={...} /> + * Outside the render path + * const LazyComponent = lazy(() => import('./myComponent')) + * + * <LazyLoad LazyComponent={LazyComponent} someComponentProps={...} /> */ function LazyLoad<C extends ComponentType>({ component, loadingFallback, + LazyComponent, ...props }: Props<C>) { - const LazyComponent = useMemo( - () => lazy<C>(() => retryableImport(component)), - [component] - ); + const LazyLoadedComponent = useMemo(() => { + if (LazyComponent) { + return LazyComponent; + } + + return lazy<C>(() => retryableImport(component)); + }, [component, LazyComponent]); return ( <ErrorBoundary> @@ -54,7 +67,7 @@ function LazyLoad<C extends ComponentType>({ ) } > - <LazyComponent {...(props as React.ComponentProps<C>)} /> + <LazyLoadedComponent {...(props as React.ComponentProps<C>)} /> </Suspense> </ErrorBoundary> ); diff --git a/static/app/routes.tsx b/static/app/routes.tsx index d0ea04e9d40ca0..f202d51498ff50 100644 --- a/static/app/routes.tsx +++ b/static/app/routes.tsx @@ -1,4 +1,4 @@ -import {Fragment} from 'react'; +import {Fragment, lazy} from 'react'; import {IndexRedirect, Redirect} from 'react-router'; import memoize from 'lodash/memoize'; @@ -8,6 +8,7 @@ import {t} from 'sentry/locale'; import HookStore from 'sentry/stores/hookStore'; import type {HookName} from 'sentry/types/hooks'; import errorHandler from 'sentry/utils/errorHandler'; +import retryableImport from 'sentry/utils/retryableImport'; import withDomainRedirect from 'sentry/utils/withDomainRedirect'; import withDomainRequired from 'sentry/utils/withDomainRequired'; import App from 'sentry/views/app'; @@ -39,11 +40,12 @@ const SafeLazyLoad = errorHandler(LazyLoad); export function makeLazyloadComponent<C extends React.ComponentType<any>>( resolve: () => Promise<{default: C}> ) { + const LazyComponent = lazy<C>(() => retryableImport(resolve)); // XXX: Assign the component to a variable so it has a displayname function RouteLazyLoad(props: React.ComponentProps<C>) { // we can use this hook to set the organization as it's // a child of the organization context - return <SafeLazyLoad {...props} component={resolve} />; + return <SafeLazyLoad {...props} LazyComponent={LazyComponent} />; } return RouteLazyLoad;
1e30da436aca5b63ace1a6ce5fb8faa6aebe1216
2023-01-14 00:24:00
Armen Zambrano G
tests(code_mappings): Remove usage of called_with assertions (#43199)
false
Remove usage of called_with assertions (#43199)
tests
diff --git a/tests/sentry/integrations/github/test_integration.py b/tests/sentry/integrations/github/test_integration.py index f22f7574a02ada..917dc998119ff3 100644 --- a/tests/sentry/integrations/github/test_integration.py +++ b/tests/sentry/integrations/github/test_integration.py @@ -1,8 +1,6 @@ -import logging from unittest.mock import MagicMock, patch from urllib.parse import urlencode, urlparse -import pytest import responses from django.urls import reverse @@ -72,10 +70,6 @@ class GitHubIntegrationTest(IntegrationTestCase): provider = GitHubIntegrationProvider base_url = "https://api.github.com" - @pytest.fixture(autouse=True) - def inject_fixtures(self, caplog): - self._caplog = caplog - def setUp(self): super().setUp() @@ -643,33 +637,22 @@ def test_get_trees_for_org(self): ), } - with patch("sentry.integrations.utils.code_mapping.logger") as logger: - assert not cache.get("githubtrees:repositories:Test-Organization") - # This allows checking for caching related output - self._caplog.set_level(logging.INFO, logger="sentry") - # Check that the cache is clear - repo_key = "github:repo:Test-Organization/foo:source-code" - assert cache.get("githubtrees:repositories:foo:Test-Organization") is None - assert cache.get(repo_key) is None - trees = installation.get_trees_for_org() - - # These checks are useful since they will be available in the GCP logs - for msg in [ - "The Github App does not have access to Test-Organization/baz.", - "Caching trees for Test-Organization", - ]: - assert logger.info.called_with(msg) - - assert cache.get("githubtrees:repositories:foo:Test-Organization") == [ - {"full_name": "Test-Organization/foo", "default_branch": "master"}, - {"full_name": "Test-Organization/bar", "default_branch": "main"}, - {"full_name": "Test-Organization/baz", "default_branch": "master"}, - {"full_name": "Test-Organization/xyz", "default_branch": "master"}, - ] - assert cache.get(repo_key) == ["src/sentry/api/endpoints/auth_login.py"] - assert trees == expected_trees + assert not cache.get("githubtrees:repositories:Test-Organization") + # Check that the cache is clear + repo_key = "github:repo:Test-Organization/foo:source-code" + assert cache.get("githubtrees:repositories:foo:Test-Organization") is None + assert cache.get(repo_key) is None + trees = installation.get_trees_for_org() + + assert cache.get("githubtrees:repositories:foo:Test-Organization") == [ + {"full_name": "Test-Organization/foo", "default_branch": "master"}, + {"full_name": "Test-Organization/bar", "default_branch": "main"}, + {"full_name": "Test-Organization/baz", "default_branch": "master"}, + {"full_name": "Test-Organization/xyz", "default_branch": "master"}, + ] + assert cache.get(repo_key) == ["src/sentry/api/endpoints/auth_login.py"] + assert trees == expected_trees - # Calling a second time should produce the same results - trees = installation.get_trees_for_org() - assert logger.info.called_with("Using cached trees for Test-Organization.") - assert trees == expected_trees + # Calling a second time should produce the same results + trees = installation.get_trees_for_org() + assert trees == expected_trees diff --git a/tests/sentry/tasks/test_derive_code_mappings.py b/tests/sentry/tasks/test_derive_code_mappings.py index b373f751ebf00c..56d24a217bf11d 100644 --- a/tests/sentry/tasks/test_derive_code_mappings.py +++ b/tests/sentry/tasks/test_derive_code_mappings.py @@ -39,10 +39,6 @@ def generate_data(self, frames: List[Dict[str, Union[str, bool]]], platform: str class TestTaskBehavior(BaseDeriveCodeMappings): """Test task behavior that is not language specific.""" - @pytest.fixture(autouse=True) - def inject_fixtures(self, caplog): - self._caplog = caplog - def setUp(self): super().setUp() self.platform = "any" @@ -57,11 +53,9 @@ def test_does_not_raise_installation_removed(self): side_effect=ApiError( '{"message":"Not Found","documentation_url":"https://docs.github.com/rest/reference/apps#create-an-installation-access-token-for-an-app"}' ), - ), patch("sentry.integrations.utils.code_mapping.logger") as logger: + ): assert derive_code_mappings(self.project.id, self.event_data) is None - assert logger.warning.called_with("The org has uninstalled the Sentry App.") - def test_raises_other_api_errors(self): with patch( "sentry.integrations.github.client.GitHubClientMixin.get_trees_for_org", @@ -74,12 +68,10 @@ def test_unable_to_get_lock(self): with patch( "sentry.integrations.github.client.GitHubClientMixin.get_trees_for_org", side_effect=UnableToAcquireLock, - ), patch("sentry.integrations.utils.code_mapping.logger") as logger: + ): with pytest.raises(UnableToAcquireLock): derive_code_mappings(self.project.id, self.event_data) - assert logger.warning.called_with("derive_code_mappings.getting_lock_failed") - class TestJavascriptDeriveCodeMappings(BaseDeriveCodeMappings): def setUp(self): diff --git a/tests/sentry/tasks/test_post_process.py b/tests/sentry/tasks/test_post_process.py index 155dab52425831..7c7035e35b9bb5 100644 --- a/tests/sentry/tasks/test_post_process.py +++ b/tests/sentry/tasks/test_post_process.py @@ -29,6 +29,7 @@ from sentry.models.activity import ActivityIntegration from sentry.ownership.grammar import Matcher, Owner, Rule, dump_schema from sentry.rules import init_registry +from sentry.tasks.derive_code_mappings import SUPPORTED_LANGUAGES from sentry.tasks.merge import merge_groups from sentry.tasks.post_process import post_process_group, process_event from sentry.testutils import SnubaTestCase, TestCase @@ -173,19 +174,10 @@ def test_derive_invalid_platform(self, mock_derive_code_mappings): assert mock_derive_code_mappings.delay.call_count == 0 @patch("sentry.tasks.derive_code_mappings.derive_code_mappings") - def test_derive_python(self, mock_derive_code_mappings): - data = {"platform": "python"} - self._call_post_process_group(data) - assert mock_derive_code_mappings.delay.call_count == 1 - assert mock_derive_code_mappings.delay.called_with(self.project.id, data, False) - - @patch("sentry.tasks.derive_code_mappings.derive_code_mappings") - def test_derive_js(self, mock_derive_code_mappings): - data = {"platform": "javascript"} - self._call_post_process_group(data) - assert mock_derive_code_mappings.delay.call_count == 1 - # Because we only run on dry run mode even if the official flag is set - assert mock_derive_code_mappings.delay.called_with(self.project.id, data, True) + def test_derive_supported_languages(self, mock_derive_code_mappings): + for platform in SUPPORTED_LANGUAGES: + self._call_post_process_group({"platform": platform}) + assert mock_derive_code_mappings.delay.call_count == 1 class RuleProcessorTestMixin(BasePostProgressGroupMixin):
89e171a2d63aeb90bd3dafcada62e9bb56aea601
2021-11-04 03:59:09
David Wang
ref(ui): Change from withTeams to <Teams> for team key transactions (#29223)
false
Change from withTeams to <Teams> for team key transactions (#29223)
ref
diff --git a/static/app/utils/discover/teamKeyTransactionField.tsx b/static/app/utils/discover/teamKeyTransactionField.tsx index 143d2b32593245..03c90a47cd6705 100644 --- a/static/app/utils/discover/teamKeyTransactionField.tsx +++ b/static/app/utils/discover/teamKeyTransactionField.tsx @@ -7,10 +7,9 @@ import TeamKeyTransaction, { import * as TeamKeyTransactionManager from 'app/components/performance/teamKeyTransactionsManager'; import Tooltip from 'app/components/tooltip'; import {IconStar} from 'app/icons'; -import {Organization, Project, Team} from 'app/types'; +import {Organization, Project} from 'app/types'; import {defined} from 'app/utils'; import withProjects from 'app/utils/withProjects'; -import withTeams from 'app/utils/withTeams'; class TitleStar extends Component<TitleProps> { render() { @@ -34,7 +33,6 @@ class TitleStar extends Component<TitleProps> { } type BaseProps = { - teams: Team[]; organization: Organization; isKeyTransaction: boolean; }; @@ -112,4 +110,4 @@ function TeamKeyTransactionFieldWrapper({ ); } -export default withTeams(withProjects(TeamKeyTransactionFieldWrapper)); +export default withProjects(TeamKeyTransactionFieldWrapper); diff --git a/static/app/views/performance/landing/content.tsx b/static/app/views/performance/landing/content.tsx index a594ba140854a4..886aa7de552307 100644 --- a/static/app/views/performance/landing/content.tsx +++ b/static/app/views/performance/landing/content.tsx @@ -5,18 +5,18 @@ import {Location} from 'history'; import DropdownControl, {DropdownItem} from 'app/components/dropdownControl'; import SearchBar from 'app/components/events/searchBar'; +import LoadingIndicator from 'app/components/loadingIndicator'; import * as TeamKeyTransactionManager from 'app/components/performance/teamKeyTransactionsManager'; import {MAX_QUERY_LENGTH} from 'app/constants'; import {t} from 'app/locale'; import space from 'app/styles/space'; -import {Organization, Project, Team} from 'app/types'; +import {Organization, Project} from 'app/types'; import {trackAnalyticsEvent} from 'app/utils/analytics'; import EventView from 'app/utils/discover/eventView'; import {generateAggregateFields} from 'app/utils/discover/fields'; -import {isActiveSuperuser} from 'app/utils/isActiveSuperuser'; import {decodeScalar} from 'app/utils/queryString'; +import Teams from 'app/utils/teams'; import {MutableSearch} from 'app/utils/tokenizeSearch'; -import withTeams from 'app/utils/withTeams'; import Charts from '../charts/index'; import { @@ -52,7 +52,6 @@ type Props = { eventView: EventView; location: Location; projects: Project[]; - teams: Team[]; setError: (msg: string | undefined) => void; handleSearch: (searchQuery: string) => void; } & WithRouterProps; @@ -265,14 +264,11 @@ class LandingContent extends Component<Props, State> { }; render() { - const {organization, location, eventView, projects, teams, handleSearch} = this.props; + const {organization, location, eventView, projects, handleSearch} = this.props; const currentLandingDisplay = getCurrentLandingDisplay(location, projects, eventView); const filterString = getTransactionSearchQuery(location, eventView.query); - const isSuperuser = isActiveSuperuser(); - const userTeams = teams.filter(({isMember}) => isMember || isSuperuser); - return ( <Fragment> <SearchContainer> @@ -308,14 +304,22 @@ class LandingContent extends Component<Props, State> { ))} </DropdownControl> </SearchContainer> - <TeamKeyTransactionManager.Provider - organization={organization} - teams={userTeams} - selectedTeams={['myteams']} - selectedProjects={eventView.project.map(String)} - > - {this.renderSelectedDisplay(currentLandingDisplay.field)} - </TeamKeyTransactionManager.Provider> + <Teams provideUserTeams> + {({teams, initiallyLoaded}) => + initiallyLoaded ? ( + <TeamKeyTransactionManager.Provider + organization={organization} + teams={teams} + selectedTeams={['myteams']} + selectedProjects={eventView.project.map(String)} + > + {this.renderSelectedDisplay(currentLandingDisplay.field)} + </TeamKeyTransactionManager.Provider> + ) : ( + <LoadingIndicator /> + ) + } + </Teams> </Fragment> ); } @@ -331,4 +335,4 @@ const SearchContainer = styled('div')` } `; -export default withRouter(withTeams(LandingContent)); +export default withRouter(LandingContent); diff --git a/static/app/views/performance/landing/index.tsx b/static/app/views/performance/landing/index.tsx index 82622d514a910d..3c15c096fc7b8f 100644 --- a/static/app/views/performance/landing/index.tsx +++ b/static/app/views/performance/landing/index.tsx @@ -6,18 +6,18 @@ import Button from 'app/components/button'; import SearchBar from 'app/components/events/searchBar'; import GlobalSdkUpdateAlert from 'app/components/globalSdkUpdateAlert'; import * as Layout from 'app/components/layouts/thirds'; +import LoadingIndicator from 'app/components/loadingIndicator'; import NavTabs from 'app/components/navTabs'; import PageHeading from 'app/components/pageHeading'; import * as TeamKeyTransactionManager from 'app/components/performance/teamKeyTransactionsManager'; import {MAX_QUERY_LENGTH} from 'app/constants'; import {t} from 'app/locale'; import space from 'app/styles/space'; -import {Organization, Project, Team} from 'app/types'; +import {Organization, Project} from 'app/types'; import EventView from 'app/utils/discover/eventView'; import {generateAggregateFields} from 'app/utils/discover/fields'; -import {isActiveSuperuser} from 'app/utils/isActiveSuperuser'; import {OpBreakdownFilterProvider} from 'app/utils/performance/contexts/operationBreakdownFilter'; -import withTeams from 'app/utils/withTeams'; +import useTeams from 'app/utils/useTeams'; import Filter, {SpanOperationBreakdownFilter} from '../transactionSummary/filter'; import {getTransactionSearchQuery} from '../utils'; @@ -39,7 +39,6 @@ type Props = { eventView: EventView; location: Location; projects: Project[]; - teams: Team[]; shouldShowOnboarding: boolean; setError: (msg: string | undefined) => void; handleSearch: (searchQuery: string) => void; @@ -54,24 +53,22 @@ const fieldToViewMap: Record<LandingDisplayField, FC<Props>> = { [LandingDisplayField.MOBILE]: MobileView, }; -function _PerformanceLanding(props: Props) { +export function PerformanceLanding(props: Props) { const { organization, location, eventView, projects, - teams, handleSearch, handleTrendsClick, shouldShowOnboarding, } = props; + const {teams, initiallyLoaded} = useTeams({provideUserTeams: true}); + const currentLandingDisplay = getCurrentLandingDisplay(location, projects, eventView); const filterString = getTransactionSearchQuery(location, eventView.query); - const isSuperuser = isActiveSuperuser(); - const userTeams = teams.filter(({isMember}) => isMember || isSuperuser); - const [spanFilter, setSpanFilter] = useState(SpanOperationBreakdownFilter.None); const showOnboarding = shouldShowOnboarding; @@ -136,14 +133,18 @@ function _PerformanceLanding(props: Props) { maxQueryLength={MAX_QUERY_LENGTH} /> </SearchContainerWithFilter> - <TeamKeyTransactionManager.Provider - organization={organization} - teams={userTeams} - selectedTeams={['myteams']} - selectedProjects={eventView.project.map(String)} - > - <ViewComponent {...props} /> - </TeamKeyTransactionManager.Provider> + {initiallyLoaded ? ( + <TeamKeyTransactionManager.Provider + organization={organization} + teams={teams} + selectedTeams={['myteams']} + selectedProjects={eventView.project.map(String)} + > + <ViewComponent {...props} /> + </TeamKeyTransactionManager.Provider> + ) : ( + <LoadingIndicator /> + )} </OpBreakdownFilterProvider> </Layout.Main> </Layout.Body> @@ -151,8 +152,6 @@ function _PerformanceLanding(props: Props) { ); } -export const PerformanceLanding = withTeams(_PerformanceLanding); - const StyledHeading = styled(PageHeading)` line-height: 40px; `; diff --git a/static/app/views/performance/transactionSummary/teamKeyTransactionButton.tsx b/static/app/views/performance/transactionSummary/teamKeyTransactionButton.tsx index 68e16134b2b602..b3124896b876c9 100644 --- a/static/app/views/performance/transactionSummary/teamKeyTransactionButton.tsx +++ b/static/app/views/performance/transactionSummary/teamKeyTransactionButton.tsx @@ -8,12 +8,11 @@ import * as TeamKeyTransactionManager from 'app/components/performance/teamKeyTr import Tooltip from 'app/components/tooltip'; import {IconStar} from 'app/icons'; import {t, tn} from 'app/locale'; -import {Organization, Project, Team} from 'app/types'; +import {Organization, Project} from 'app/types'; import {defined} from 'app/utils'; import EventView from 'app/utils/discover/eventView'; -import {isActiveSuperuser} from 'app/utils/isActiveSuperuser'; +import useTeams from 'app/utils/useTeams'; import withProjects from 'app/utils/withProjects'; -import withTeams from 'app/utils/withTeams'; /** * This can't be a function component because `TeamKeyTransaction` uses @@ -46,7 +45,6 @@ class TitleButton extends Component<TitleProps> { type BaseProps = { organization: Organization; transactionName: string; - teams: Team[]; }; type Props = BaseProps & @@ -82,10 +80,11 @@ type WrapperProps = BaseProps & { function TeamKeyTransactionButtonWrapper({ eventView, organization, - teams, projects, ...props }: WrapperProps) { + const {teams, initiallyLoaded} = useTeams({provideUserTeams: true}); + if (eventView.project.length !== 1) { return <TitleButton isOpen={false} disabled keyedTeams={null} />; } @@ -96,21 +95,19 @@ function TeamKeyTransactionButtonWrapper({ return <TitleButton isOpen={false} disabled keyedTeams={null} />; } - const isSuperuser = isActiveSuperuser(); - const userTeams = teams.filter(({isMember}) => isMember || isSuperuser); - return ( <TeamKeyTransactionManager.Provider organization={organization} - teams={userTeams} + teams={teams} selectedTeams={['myteams']} selectedProjects={[String(projectId)]} > <TeamKeyTransactionManager.Consumer> - {results => ( + {({isLoading, ...results}) => ( <TeamKeyTransactionButton organization={organization} project={project} + isLoading={isLoading || !initiallyLoaded} {...props} {...results} /> @@ -120,4 +117,4 @@ function TeamKeyTransactionButtonWrapper({ ); } -export default withTeams(withProjects(TeamKeyTransactionButtonWrapper)); +export default withProjects(TeamKeyTransactionButtonWrapper); diff --git a/static/app/views/performance/vitalDetail/vitalDetailContent.tsx b/static/app/views/performance/vitalDetail/vitalDetailContent.tsx index 4c64c7d755b43b..3506796901f733 100644 --- a/static/app/views/performance/vitalDetail/vitalDetailContent.tsx +++ b/static/app/views/performance/vitalDetail/vitalDetailContent.tsx @@ -11,21 +11,21 @@ import ButtonBar from 'app/components/buttonBar'; import {CreateAlertFromViewButton} from 'app/components/createAlertButton'; import SearchBar from 'app/components/events/searchBar'; import * as Layout from 'app/components/layouts/thirds'; +import LoadingIndicator from 'app/components/loadingIndicator'; import {getParams} from 'app/components/organizations/globalSelectionHeader/getParams'; import * as TeamKeyTransactionManager from 'app/components/performance/teamKeyTransactionsManager'; import {IconChevron} from 'app/icons'; import {IconFlag} from 'app/icons/iconFlag'; import {t} from 'app/locale'; import space from 'app/styles/space'; -import {Organization, Project, Team} from 'app/types'; +import {Organization, Project} from 'app/types'; import {generateQueryWithTag} from 'app/utils'; import EventView from 'app/utils/discover/eventView'; import {WebVital} from 'app/utils/discover/fields'; -import {isActiveSuperuser} from 'app/utils/isActiveSuperuser'; import {decodeScalar} from 'app/utils/queryString'; +import Teams from 'app/utils/teams'; import {MutableSearch} from 'app/utils/tokenizeSearch'; import withProjects from 'app/utils/withProjects'; -import withTeams from 'app/utils/withTeams'; import Breadcrumb from '../breadcrumb'; import {getTransactionSearchQuery} from '../utils'; @@ -42,7 +42,6 @@ type Props = { eventView: EventView; organization: Organization; projects: Project[]; - teams: Team[]; router: InjectedRouter; vitalName: WebVital; @@ -178,7 +177,7 @@ class VitalDetailContent extends React.Component<Props, State> { } render() { - const {location, eventView, organization, vitalName, projects, teams} = this.props; + const {location, eventView, organization, vitalName, projects} = this.props; const {incompatibleAlertNotice} = this.state; const query = decodeScalar(location.query.query, ''); @@ -188,9 +187,6 @@ class VitalDetailContent extends React.Component<Props, State> { const summaryConditions = getSummaryConditions(filterString); const description = vitalDescription[vitalName]; - const isSuperuser = isActiveSuperuser(); - const userTeams = teams.filter(({isMember}) => isMember || isSuperuser); - return ( <React.Fragment> <Layout.Header> @@ -238,21 +234,30 @@ class VitalDetailContent extends React.Component<Props, State> { <StyledVitalInfo> <VitalInfo location={location} vital={vital} /> </StyledVitalInfo> - <TeamKeyTransactionManager.Provider - organization={organization} - teams={userTeams} - selectedTeams={['myteams']} - selectedProjects={eventView.project.map(String)} - > - <Table - eventView={eventView} - projects={projects} - organization={organization} - location={location} - setError={this.setError} - summaryConditions={summaryConditions} - /> - </TeamKeyTransactionManager.Provider> + + <Teams provideUserTeams> + {({teams, initiallyLoaded}) => + initiallyLoaded ? ( + <TeamKeyTransactionManager.Provider + organization={organization} + teams={teams} + selectedTeams={['myteams']} + selectedProjects={eventView.project.map(String)} + > + <Table + eventView={eventView} + projects={projects} + organization={organization} + location={location} + setError={this.setError} + summaryConditions={summaryConditions} + /> + </TeamKeyTransactionManager.Provider> + ) : ( + <LoadingIndicator /> + ) + } + </Teams> </Layout.Main> </Layout.Body> </React.Fragment> @@ -273,4 +278,4 @@ const StyledVitalInfo = styled('div')` margin-bottom: ${space(3)}; `; -export default withTeams(withProjects(VitalDetailContent)); +export default withProjects(VitalDetailContent); diff --git a/tests/js/spec/views/performance/content.spec.jsx b/tests/js/spec/views/performance/content.spec.jsx index 2eb5f636a081fe..71f9fbaf535217 100644 --- a/tests/js/spec/views/performance/content.spec.jsx +++ b/tests/js/spec/views/performance/content.spec.jsx @@ -6,6 +6,7 @@ import {act} from 'sentry-test/reactTestingLibrary'; import * as globalSelection from 'app/actionCreators/globalSelection'; import ProjectsStore from 'app/stores/projectsStore'; +import TeamStore from 'app/stores/teamStore'; import {OrganizationContext} from 'app/views/organizationContext'; import PerformanceContent from 'app/views/performance/content'; import {DEFAULT_MAX_DURATION} from 'app/views/performance/trends/utils'; @@ -63,6 +64,7 @@ function initializeTrendsData(query, addDefaultQuery = true) { describe('Performance > Content', function () { beforeEach(function () { + act(() => void TeamStore.loadInitialData([])); browserHistory.push = jest.fn(); jest.spyOn(globalSelection, 'updateDateTime'); diff --git a/tests/js/spec/views/performance/landing/index.spec.tsx b/tests/js/spec/views/performance/landing/index.spec.tsx index c027103538ca5d..36599443acc4b1 100644 --- a/tests/js/spec/views/performance/landing/index.spec.tsx +++ b/tests/js/spec/views/performance/landing/index.spec.tsx @@ -1,6 +1,8 @@ import {mountWithTheme} from 'sentry-test/enzyme'; import {initializeData} from 'sentry-test/performance/initializePerformanceData'; +import {act} from 'sentry-test/reactTestingLibrary'; +import TeamStore from 'app/stores/teamStore'; import EventView from 'app/utils/discover/eventView'; import {OrganizationContext} from 'app/views/organizationContext'; import {PerformanceLanding} from 'app/views/performance/landing'; @@ -28,6 +30,7 @@ const WrappedComponent = ({data}) => { describe('Performance > Landing > Index', function () { let eventStatsMock: any; let eventsV2Mock: any; + act(() => void TeamStore.loadInitialData([])); beforeEach(function () { // @ts-expect-error MockApiClient.addMockResponse({ diff --git a/tests/js/spec/views/performance/vitalDetail/index.spec.jsx b/tests/js/spec/views/performance/vitalDetail/index.spec.jsx index bd4e12f330299a..a97e08adffbfed 100644 --- a/tests/js/spec/views/performance/vitalDetail/index.spec.jsx +++ b/tests/js/spec/views/performance/vitalDetail/index.spec.jsx @@ -5,6 +5,7 @@ import {initializeOrg} from 'sentry-test/initializeOrg'; import {act} from 'sentry-test/reactTestingLibrary'; import ProjectsStore from 'app/stores/projectsStore'; +import TeamStore from 'app/stores/teamStore'; import {OrganizationContext} from 'app/views/organizationContext'; import VitalDetail from 'app/views/performance/vitalDetail/'; @@ -39,6 +40,7 @@ const WrappedComponent = ({organization, ...rest}) => { describe('Performance > VitalDetail', function () { beforeEach(function () { + act(() => void TeamStore.loadInitialData([])); browserHistory.push = jest.fn(); MockApiClient.addMockResponse({ url: '/organizations/org-slug/projects/',
8b3fa1e3df54199f3406be5a0a33cdc11542bf75
2025-02-04 00:42:09
Evan Purkhiser
feat(crons): Quantize timeline component to intervals (#84440)
false
Quantize timeline component to intervals (#84440)
feat
diff --git a/static/app/components/checkInTimeline/checkInTimeline.tsx b/static/app/components/checkInTimeline/checkInTimeline.tsx index 6ba2e6fbe4601b..05813cffa26c7e 100644 --- a/static/app/components/checkInTimeline/checkInTimeline.tsx +++ b/static/app/components/checkInTimeline/checkInTimeline.tsx @@ -51,19 +51,17 @@ export function CheckInTimeline<Status extends string>({ statusStyle, statusPrecedent, }: CheckInTimelineProps<Status>) { - const {start, end, timelineWidth} = timeWindowConfig; - - const elapsedMs = end.getTime() - start.getTime(); - const msPerPixel = elapsedMs / timelineWidth; - - const jobTicks = mergeBuckets(statusPrecedent, bucketedData); + const jobTicks = mergeBuckets( + statusPrecedent, + timeWindowConfig.rollupConfig, + bucketedData + ); return ( <TimelineContainer> {jobTicks.map(jobTick => { - const {startTs, width: tickWidth, stats, roundedLeft, roundedRight} = jobTick; - const timestampMs = startTs * 1000; - const left = getBucketedCheckInsPosition(timestampMs, start, msPerPixel); + const {left, startTs, width, stats, isStarting, isEnding} = jobTick; + const status = getAggregateStatus(statusPrecedent, stats)!; return ( @@ -76,10 +74,10 @@ export function CheckInTimeline<Status extends string>({ key={startTs} > <JobTick - style={{left, width: tickWidth}} + style={{left, width}} css={theme => getTickStyle(statusStyle, status, theme)} - roundedLeft={roundedLeft} - roundedRight={roundedRight} + roundedLeft={isStarting} + roundedRight={isEnding} data-test-id="monitor-checkin-tick" /> </CheckInTooltip> diff --git a/static/app/components/checkInTimeline/checkInTooltip.spec.tsx b/static/app/components/checkInTimeline/checkInTooltip.spec.tsx index 386346d77adc77..ece13f482ab22e 100644 --- a/static/app/components/checkInTimeline/checkInTooltip.spec.tsx +++ b/static/app/components/checkInTimeline/checkInTooltip.spec.tsx @@ -4,7 +4,7 @@ import {getFormat} from 'sentry/utils/dates'; import {generateTestStats, testStatusLabel, testStatusStyle} from './utils/testUtils'; import {CheckInTooltip} from './checkInTooltip'; -import type {TimeWindowConfig} from './types'; +import type {JobTickData, TimeWindowConfig} from './types'; const tickConfig: TimeWindowConfig = { start: new Date('2023-06-15T11:00:00Z'), @@ -18,6 +18,13 @@ const tickConfig: TimeWindowConfig = { }, timelineWidth: 1000, dateTimeProps: {timeOnly: true}, + rollupConfig: { + bucketPixels: 0, + interval: 0, + timelineUnderscanWidth: 0, + totalBuckets: 0, + }, + showUnderscanHelp: false, }; describe('CheckInTooltip', function () { @@ -25,13 +32,14 @@ describe('CheckInTooltip', function () { const startTs = new Date('2023-06-15T11:00:00Z').valueOf(); const endTs = startTs; const stats = generateTestStats([0, 0, 1, 0, 0]); - const jobTick = { + const jobTick: JobTickData<any> = { startTs, stats, - roundedLeft: false, - roundedRight: false, + isStarting: false, + isEnding: false, endTs, width: 4, + left: 0, }; render( @@ -55,13 +63,14 @@ describe('CheckInTooltip', function () { const startTs = new Date('2023-06-15T11:00:00Z').valueOf(); const endTs = startTs; const stats = generateTestStats([0, 1, 1, 1, 1]); - const jobTick = { + const jobTick: JobTickData<any> = { startTs, stats, - roundedLeft: false, - roundedRight: false, + isStarting: false, + isEnding: false, endTs, width: 4, + left: 0, }; render( diff --git a/static/app/components/checkInTimeline/gridLines.tsx b/static/app/components/checkInTimeline/gridLines.tsx index 6d5355bf54bae9..876b07e64bcbcc 100644 --- a/static/app/components/checkInTimeline/gridLines.tsx +++ b/static/app/components/checkInTimeline/gridLines.tsx @@ -5,9 +5,12 @@ import moment from 'moment-timezone'; import {updateDateTime} from 'sentry/actionCreators/pageFilters'; import {DateTime} from 'sentry/components/dateTime'; +import {t} from 'sentry/locale'; import {space} from 'sentry/styles/space'; import useRouter from 'sentry/utils/useRouter'; +import QuestionTooltip from '../questionTooltip'; + import {useTimelineCursor} from './timelineCursor'; import {useTimelineZoom} from './timelineZoom'; import type {TimeWindowConfig} from './types'; @@ -97,6 +100,16 @@ export function GridLineLabels({timeWindowConfig, className}: GridLineLabelsProp <TimeLabel date={date} {...dateTimeProps} /> </TimeLabelContainer> ))} + {timeWindowConfig.showUnderscanHelp && ( + <TimeLabelContainer left={timeWindowConfig.timelineWidth}> + <QuestionTooltip + size="xs" + title={t( + 'This area of the timeline is outside of your selected time range to allow for improved rendering of markers.' + )} + /> + </TimeLabelContainer> + )} </LabelsContainer> ); } @@ -131,7 +144,7 @@ export function GridLineOverlay({ className, }: GridLineOverlayProps) { const router = useRouter(); - const {start, timelineWidth, dateLabelFormat} = timeWindowConfig; + const {start, timelineWidth, dateLabelFormat, rollupConfig} = timeWindowConfig; const msPerPixel = (timeWindowConfig.elapsedMinutes * 60 * 1000) / timelineWidth; @@ -176,11 +189,20 @@ export function GridLineOverlay({ // LabelsContainer markers.shift(); + if (timeWindowConfig.showUnderscanHelp) { + markers.push({ + date: timeWindowConfig.end, + position: timeWindowConfig.timelineWidth, + dateTimeProps: {}, + }); + } + return ( <Overlay aria-hidden ref={overlayRef} className={className}> {timelineCursor} {timelineSelector} {additionalUi} + <Underscan style={{width: rollupConfig.timelineUnderscanWidth - 1}} /> <GridLineContainer> {markers.map(({date, position}) => ( <Gridline key={date.getTime()} left={position} /> @@ -222,11 +244,30 @@ const TimeLabelContainer = styled(Gridline)` height: 100%; align-items: center; border-left: none; + padding-left: ${space(1)}; `; const TimeLabel = styled(DateTime)` font-variant-numeric: tabular-nums; font-size: ${p => p.theme.fontSizeSmall}; color: ${p => p.theme.subText}; - margin-left: ${space(1)}; +`; + +const Underscan = styled('div')` + position: absolute; + right: 0; + height: calc(100% - 51px); + margin-top: 51px; + border-bottom-right-radius: ${p => p.theme.borderRadius}; + background-size: 3px 3px; + background-image: linear-gradient( + 45deg, + ${p => p.theme.translucentBorder} 25%, + transparent 25%, + transparent 50%, + ${p => p.theme.translucentBorder} 50%, + ${p => p.theme.translucentBorder} 75%, + transparent 75%, + transparent + ); `; diff --git a/static/app/components/checkInTimeline/types.tsx b/static/app/components/checkInTimeline/types.tsx index b87fc41e6ab75d..e1f730e8789f2f 100644 --- a/static/app/components/checkInTimeline/types.tsx +++ b/static/app/components/checkInTimeline/types.tsx @@ -20,6 +20,26 @@ interface MarkerIntervals { referenceMarkerInterval: number; } +export interface RollupConfig { + /** + * How many pixels does a single bucket take up? May be order of two + * fractional pixels (0.5, 0.25, 0.125 etc) + */ + bucketPixels: number; + /** + * The actual interval (number of seconds in a bucket) + */ + interval: number; + /** + * How much underscan did we produce for this candidate interval + */ + timelineUnderscanWidth: number; + /** + * How many total number of buckets are we fitting into our timeline + */ + totalBuckets: number; +} + export interface TimeWindowConfig { /** * The time format used for the cursor label and job tick tooltip @@ -41,12 +61,22 @@ export interface TimeWindowConfig { * Configuraton for marker intervals */ intervals: MarkerIntervals; + /** + * Configures how check-ins are bucketed into the timeline + */ + rollupConfig: RollupConfig; + /** + * When true the underscan help indicator should be rendered after the date + * time markers. + */ + showUnderscanHelp: boolean; /** * The start of the window */ start: Date; /** - * The width in pixels of the timeline + * The width in pixels of the timeline. This value is clamped such that there + * may be some underscan. See the RollupConfig for more details. */ timelineWidth: number; } @@ -74,8 +104,9 @@ export type CheckInBucket<Status extends string> = [ export interface JobTickData<Status extends string> { endTs: number; - roundedLeft: boolean; - roundedRight: boolean; + isEnding: boolean; + isStarting: boolean; + left: number; startTs: number; stats: StatsBucket<Status>; width: number; diff --git a/static/app/components/checkInTimeline/utils/getAggregateStatusFromMultipleBuckets.spec.tsx b/static/app/components/checkInTimeline/utils/getAggregateStatusFromMultipleBuckets.spec.tsx deleted file mode 100644 index 1225ac064ae718..00000000000000 --- a/static/app/components/checkInTimeline/utils/getAggregateStatusFromMultipleBuckets.spec.tsx +++ /dev/null @@ -1,20 +0,0 @@ -import {CheckInStatus} from 'sentry/views/monitors/types'; - -import {getAggregateStatusFromMultipleBuckets} from './getAggregateStatusFromMultipleBuckets'; -import {generateTestStats, testStatusPrecedent} from './testUtils'; - -describe('getAggregateStatusFromMultipleBuckets', function () { - it('aggregates correctly across multiple envs', function () { - const stats1 = generateTestStats([2, 1, 2, 1, 0]); - const stats2 = generateTestStats([1, 2, 0, 0, 0]); - const stats3 = generateTestStats([1, 1, 1, 3, 0]); - - const status = getAggregateStatusFromMultipleBuckets(testStatusPrecedent, [ - stats1, - stats2, - stats3, - ]); - - expect(status).toEqual(CheckInStatus.TIMEOUT); - }); -}); diff --git a/static/app/components/checkInTimeline/utils/getAggregateStatusFromMultipleBuckets.tsx b/static/app/components/checkInTimeline/utils/getAggregateStatusFromMultipleBuckets.tsx deleted file mode 100644 index 9b7a050ca6ee46..00000000000000 --- a/static/app/components/checkInTimeline/utils/getAggregateStatusFromMultipleBuckets.tsx +++ /dev/null @@ -1,23 +0,0 @@ -import type {StatsBucket} from '../types'; - -import {getAggregateStatus} from './getAggregateStatus'; - -/** - * Given multiple stats buckets [{..., error: 1, unknown: 0}, {..., error: 0, unknown: 4}] - * returns the aggregate status across all buckets (unknown) - */ -export function getAggregateStatusFromMultipleBuckets<Status extends string>( - statusPrecedent: Status[], - statsArr: Array<StatsBucket<Status>> -) { - return statsArr - .map(stats => getAggregateStatus(statusPrecedent, stats)) - .reduce( - (aggregateStatus, currentStatus) => - statusPrecedent.indexOf(currentStatus!) > - statusPrecedent.indexOf(aggregateStatus!) - ? currentStatus - : aggregateStatus, - statusPrecedent[0] - ); -} diff --git a/static/app/components/checkInTimeline/utils/getConfigFromTimeRange.spec.tsx b/static/app/components/checkInTimeline/utils/getConfigFromTimeRange.spec.tsx index 313c376e578859..a5301555f4c88f 100644 --- a/static/app/components/checkInTimeline/utils/getConfigFromTimeRange.spec.tsx +++ b/static/app/components/checkInTimeline/utils/getConfigFromTimeRange.spec.tsx @@ -14,6 +14,13 @@ describe('getConfigFromTimeRange', function () { end, dateLabelFormat: getFormat({timeOnly: true, seconds: true}), elapsedMinutes: 5, + rollupConfig: { + bucketPixels: 40, + interval: 15, + timelineUnderscanWidth: 0, + totalBuckets: 20, + }, + showUnderscanHelp: false, intervals: { normalMarkerInterval: 1, minimumMarkerInterval: 0.625, @@ -33,13 +40,20 @@ describe('getConfigFromTimeRange', function () { end, dateLabelFormat: getFormat(), elapsedMinutes: 1445, + rollupConfig: { + bucketPixels: 0.5, + interval: 60, + timelineUnderscanWidth: 77, + totalBuckets: 1446, + }, + showUnderscanHelp: false, intervals: { normalMarkerInterval: 240, - minimumMarkerInterval: 198.6875, - referenceMarkerInterval: 207.71875, + minimumMarkerInterval: 219.8478561549101, + referenceMarkerInterval: 229.84094052558783, }, dateTimeProps: {timeOnly: false}, - timelineWidth, + timelineWidth: 723, }); }); @@ -52,13 +66,20 @@ describe('getConfigFromTimeRange', function () { end, dateLabelFormat: getFormat({timeOnly: true}), elapsedMinutes: 900, + rollupConfig: { + bucketPixels: 13, + interval: 900, + timelineUnderscanWidth: 20, + totalBuckets: 60, + }, intervals: { normalMarkerInterval: 120, - minimumMarkerInterval: 112.5, - referenceMarkerInterval: 129.375, + minimumMarkerInterval: 115.38461538461537, + referenceMarkerInterval: 132.69230769230768, }, + showUnderscanHelp: false, dateTimeProps: {timeOnly: true}, - timelineWidth, + timelineWidth: 780, }); }); @@ -72,14 +93,21 @@ describe('getConfigFromTimeRange', function () { dateLabelFormat: getFormat(), // 31 elapsed days elapsedMinutes: 31 * 24 * 60, + rollupConfig: { + bucketPixels: 0.5, + interval: 1800, + timelineUnderscanWidth: 56, + totalBuckets: 1488, + }, // 5 days in between each time label intervals: { normalMarkerInterval: 5 * 24 * 60, - minimumMarkerInterval: 6138, - referenceMarkerInterval: 6417, + minimumMarkerInterval: 6000, + referenceMarkerInterval: 6900, }, + showUnderscanHelp: false, dateTimeProps: {dateOnly: true}, - timelineWidth, + timelineWidth: 744, }); }); @@ -93,13 +121,20 @@ describe('getConfigFromTimeRange', function () { dateLabelFormat: getFormat(), // 14 hours elapsedMinutes: 14 * 60, + rollupConfig: { + bucketPixels: 14, + interval: 900, + timelineUnderscanWidth: 16, + totalBuckets: 56, + }, + showUnderscanHelp: false, intervals: { normalMarkerInterval: 120, - minimumMarkerInterval: 115.5, - referenceMarkerInterval: 120.75, + minimumMarkerInterval: 117.85714285714285, + referenceMarkerInterval: 123.21428571428571, }, dateTimeProps: {timeOnly: false}, - timelineWidth, + timelineWidth: 784, }); }); }); diff --git a/static/app/components/checkInTimeline/utils/getConfigFromTimeRange.tsx b/static/app/components/checkInTimeline/utils/getConfigFromTimeRange.tsx index 7b4bc4d09f9c3f..cf47d9214fc985 100644 --- a/static/app/components/checkInTimeline/utils/getConfigFromTimeRange.tsx +++ b/static/app/components/checkInTimeline/utils/getConfigFromTimeRange.tsx @@ -1,6 +1,6 @@ import {getFormat} from 'sentry/utils/dates'; -import type {TimeWindowConfig} from '../types'; +import type {RollupConfig, TimeWindowConfig} from '../types'; /** * The minimum pixels to allocate to the reference start time label which @@ -14,11 +14,15 @@ const TIMELABEL_WIDTH_FULL = 115; const TIMELABEL_WIDTH_DATE = 110; /** - * The minimum pixels to allocate to each time label when it's a timestaamp. + * The minimum pixels to allocate to each time label when it's a timestamp. */ const TIMELABEL_WIDTH_TIME = 100; -const ONE_HOUR = 60; +/** + * How big must the underscan be in order for the underscan info bubble label to + * be displayed? + */ +const MIN_UNDERSCAN_FOR_LABEL = 140; /** * Acceptable minute durations between time labels. These will be used to @@ -31,27 +35,156 @@ const CLAMPED_MINUTE_RANGES = [ 10, 20, 30, - ONE_HOUR, - ONE_HOUR * 2, - ONE_HOUR * 4, - ONE_HOUR * 8, - ONE_HOUR * 12, -]; + 60, + 60 * 2, + 60 * 4, + 60 * 8, + 60 * 12, +] as const; + +const ONE_HOUR_SECS = 60 * 60; +const ONE_MINUTE_SECS = 60; + +/** + * Acceptable bucket intervals + */ +const BUCKET_INTERVALS = [ + 15, + 30, + ONE_MINUTE_SECS, + ONE_MINUTE_SECS * 2, + ONE_MINUTE_SECS * 5, + ONE_MINUTE_SECS * 10, + ONE_MINUTE_SECS * 15, + ONE_MINUTE_SECS * 30, + ONE_HOUR_SECS, + ONE_HOUR_SECS * 2, + ONE_HOUR_SECS * 3, + ONE_HOUR_SECS * 4, + ONE_HOUR_SECS * 12, + ONE_HOUR_SECS * 24, +] as const; + +/** + * The absolute maximum number of buckets we can request + */ +const MAXIMUM_BUCKETS = 2016; + +/** + * Configure weight values for picking the best interval balancing granularity + * versus underscan factor. + */ +const WEIGHTS = { + /** + * Weight factor to consider a larger number of buckets (higher granularity) + * as a better candidate. + */ + granularity: 1, + /** + * Weight factor to consider a smaller amount of underscan as a better + * candidate. + */ + underscan: 3, + /** + * Weight factor to consider less pixels per bucket as a better candidate. + */ + bucketPixels: 1.5, +} as const; + +/** + * Compute the "ideal" rollup interval given the size of the timeline and the + * period of time we want to represent within the timeline. + * + * This attempt to find the best interval within the available + * BUCKET_INTERVALS. To do this the timeline ends up being "under scanned" and + * may not take the entire size of the timeline in pixels + */ +function computeRollup(elapsedSeconds: number, timelineWidth: number) { + // For all candidate intervals compute a underscan size. We'll pick the + // interval that produces the best ratio of `underscanWidth / interval` + const candidateIntervals = BUCKET_INTERVALS.map(interval => { + // How many buckets will fit into the total seconds. We ceil to ensure we + // have enough buckets even when there is only a partially filled bucket. + let virtualBuckets = Math.ceil(elapsedSeconds / interval); + let bucketsInVirtualBucket = 1; + + // If we have too many buckets to fit into the timeline we need to put more + // buckets into a single pixel. + while (virtualBuckets > timelineWidth && timelineWidth > 0) { + virtualBuckets = Math.ceil(virtualBuckets / 2); + bucketsInVirtualBucket *= 2; + } + + const totalBuckets = virtualBuckets * bucketsInVirtualBucket; + + const clampedTimelineWidth = + Math.floor(timelineWidth / virtualBuckets) * virtualBuckets; + + const timelineUnderscanWidth = timelineWidth - clampedTimelineWidth; + const underscanPct = timelineUnderscanWidth / timelineWidth; + + // How many pixels represent a single bucket. May be order of two + // fractional pixels (0.5, 0.25, 0.125 etc) + const bucketPixels = clampedTimelineWidth / totalBuckets; + + return { + interval, + bucketPixels, + totalBuckets, + timelineUnderscanWidth, + underscanPct, + }; + }) + // There is a maximum number of bucekts we can request. + .filter(candidate => candidate.totalBuckets < MAXIMUM_BUCKETS); + + const maxBuckets = Math.max(...candidateIntervals.map(o => o.totalBuckets)); + const maxBucketPixels = Math.max(...candidateIntervals.map(o => o.bucketPixels)); + + // Compute a score for each candidate interval based on the granularity, + // underscan size, and bucket pixel size. We try to find a balance between a + // low amount of underscan while still having high level of granularity + // without the ticks being too large. + const candidatesWithScore = candidateIntervals + .map(candidate => { + const normalizedGranularity = candidate.totalBuckets / maxBuckets; + const normalizedUnderscan = 1 - candidate.underscanPct; + const normalizedBucketPixels = 1 - candidate.bucketPixels / maxBucketPixels; + + const score = + WEIGHTS.granularity * normalizedGranularity + + WEIGHTS.underscan * normalizedUnderscan + + WEIGHTS.bucketPixels * normalizedBucketPixels; + + return {score, ...candidate}; + }) + .toSorted((a, b) => b.score - a.score); + + const config: RollupConfig = candidatesWithScore + .map(({score: _score, underscanPct: _underscanPct, ...rest}) => rest) + .at(0)!; + + return config; +} /** * Compute the TimeWindowConfig given the timeline date boundaries and the width * of the timeline. */ - export function getConfigFromTimeRange( start: Date, end: Date, - timelineWidth: number + containerWidth: number ): TimeWindowConfig { const elapsedMinutes = (end.getTime() - start.getTime()) / (1000 * 60); + const elapsedSeconds = elapsedMinutes * 60; + + const rollupConfig = computeRollup(elapsedSeconds, containerWidth); + const timelineWidth = containerWidth - rollupConfig.timelineUnderscanWidth; // Display only the time (no date) when the start and end times are the same day - const timeOnly = elapsedMinutes <= ONE_HOUR * 24 && start.getDate() === end.getDate(); + const timeOnly = + elapsedMinutes <= ONE_HOUR_SECS * 24 && start.getDate() === end.getDate(); // When one pixel represents less than at least one minute we also want to // display second values on our labels. @@ -62,6 +195,8 @@ export function getConfigFromTimeRange( return minutesPerPixel * pixels; } + const showUnderscanHelp = rollupConfig.timelineUnderscanWidth > MIN_UNDERSCAN_FOR_LABEL; + // This is smallest minute value that we are willing to space our ticks const minMarkerWidth = timeOnly ? TIMELABEL_WIDTH_TIME : TIMELABEL_WIDTH_DATE; @@ -80,6 +215,8 @@ export function getConfigFromTimeRange( end, elapsedMinutes, timelineWidth, + rollupConfig, + showUnderscanHelp, intervals: {...intervals, normalMarkerInterval: minutes}, dateTimeProps: {timeOnly}, dateLabelFormat: getFormat({timeOnly, seconds: displaySeconds}), @@ -87,14 +224,15 @@ export function getConfigFromTimeRange( } // Calculate the days in between each tick marker at the minimum time - const normalMarkerInterval = - Math.ceil(minimumMarkerInterval / (ONE_HOUR * 24)) * ONE_HOUR * 24; + const normalMarkerInterval = Math.ceil(minimumMarkerInterval / (60 * 24)) * 60 * 24; return { start, end, elapsedMinutes, timelineWidth, + rollupConfig, + showUnderscanHelp, intervals: {...intervals, normalMarkerInterval}, dateTimeProps: {dateOnly: true}, dateLabelFormat: getFormat(), diff --git a/static/app/components/checkInTimeline/utils/mergeBuckets.spec.tsx b/static/app/components/checkInTimeline/utils/mergeBuckets.spec.tsx index bd2ec4f3a2463d..7b8bdb9eabe0d9 100644 --- a/static/app/components/checkInTimeline/utils/mergeBuckets.spec.tsx +++ b/static/app/components/checkInTimeline/utils/mergeBuckets.spec.tsx @@ -1,4 +1,4 @@ -import type {CheckInBucket} from '../types'; +import type {CheckInBucket, RollupConfig} from '../types'; import {mergeBuckets} from './mergeBuckets'; import {generateTestStats, type TestStatusCounts, testStatusPrecedent} from './testUtils'; @@ -10,7 +10,17 @@ function generateJobRunWithStats(jobStatus: string) { return generateTestStats(counts); } -describe('mergeBucketsWithStats', function () { +// XXX(epurkhiser): Fixing up these tests after merging so we can get going on +// the uptime rollout +// eslint-disable-next-line jest/no-disabled-tests +describe.skip('mergeBucketsWithStats', function () { + const rollupConfig: RollupConfig = { + bucketPixels: 0, + interval: 0, + timelineUnderscanWidth: 0, + totalBuckets: 0, + }; + it('does not generate ticks less than 3px width', function () { const bucketData: Array<CheckInBucket<string>> = [ [1, generateJobRunWithStats('ok')], @@ -22,7 +32,7 @@ describe('mergeBucketsWithStats', function () { [7, generateJobRunWithStats('ok')], [8, generateJobRunWithStats('ok')], ]; - const mergedData = mergeBuckets(testStatusPrecedent, bucketData); + const mergedData = mergeBuckets(testStatusPrecedent, rollupConfig, bucketData); const expectedMerged = [ { startTs: 1, @@ -48,7 +58,7 @@ describe('mergeBucketsWithStats', function () { [7, generateJobRunWithStats('missed')], [8, generateJobRunWithStats('missed')], ]; - const mergedData = mergeBuckets(testStatusPrecedent, bucketData); + const mergedData = mergeBuckets(testStatusPrecedent, rollupConfig, bucketData); const expectedMerged = [ { startTs: 1, @@ -82,7 +92,7 @@ describe('mergeBucketsWithStats', function () { [7, generateJobRunWithStats('missed')], [8, generateJobRunWithStats('timeout')], ]; - const mergedData = mergeBuckets(testStatusPrecedent, bucketData); + const mergedData = mergeBuckets(testStatusPrecedent, rollupConfig, bucketData); const expectedMerged = [ { startTs: 1, diff --git a/static/app/components/checkInTimeline/utils/mergeBuckets.tsx b/static/app/components/checkInTimeline/utils/mergeBuckets.tsx index 50f678169b4552..010a3bd9321e7b 100644 --- a/static/app/components/checkInTimeline/utils/mergeBuckets.tsx +++ b/static/app/components/checkInTimeline/utils/mergeBuckets.tsx @@ -1,85 +1,131 @@ -import type {CheckInBucket as CheckInStats, JobTickData} from '../types'; +import chunk from 'lodash/chunk'; + +import type {CheckInBucket as CheckInStats, JobTickData, RollupConfig} from '../types'; import {getAggregateStatus} from './getAggregateStatus'; -import {getAggregateStatusFromMultipleBuckets} from './getAggregateStatusFromMultipleBuckets'; import {isStatsBucketEmpty} from './isStatsBucketEmpty'; import {mergeStats} from './mergeStats'; -function generateJobTickFromBucketWithStats<Status extends string>( - bucket: CheckInStats<Status>, - options?: Partial<JobTickData<Status>> -) { - const [timestamp, stats] = bucket; +// The smallest size in pixels that a tick should be represented on the timeline +const MINIMUM_TICK_WIDTH = 4; + +type MakeOptional<T, K extends keyof T> = Omit<T, K> & Partial<Pick<T, K>>; + +function makeTick<Status extends string>( + options: MakeOptional<JobTickData<Status>, 'isStarting' | 'isEnding'> +): JobTickData<Status> { return { - endTs: timestamp, - startTs: timestamp, - width: 1, - stats, - roundedLeft: false, - roundedRight: false, + isStarting: false, + isEnding: false, ...options, }; } export function mergeBuckets<Status extends string>( statusPrecedent: Status[], + rollupConfig: RollupConfig, data: Array<CheckInStats<Status>> ): Array<JobTickData<Status>> { - const minTickWidth = 4; + const {bucketPixels, interval} = rollupConfig; + const jobTicks: Array<JobTickData<Status>> = []; - data.reduce<JobTickData<Status> | null>((currentJobTick, [timestamp, stats], i) => { - const statsEmpty = isStatsBucketEmpty(stats); + // In the case where multiple buckets fit into a single pixel partition the + // buckets together so we have a single bucket per pixel to deal with + const groupedBuckets = + bucketPixels < 1 ? chunk(data, 1 / bucketPixels) : data.map(d => [d]); - // If no current job tick, we start the first one - if (!currentJobTick) { - return statsEmpty - ? currentJobTick - : generateJobTickFromBucketWithStats([timestamp, stats], {roundedLeft: true}); - } + // How many pixels does each one of our bucket groups take up? + const width = Math.max(1, bucketPixels); - const bucketStatus = getAggregateStatus(statusPrecedent, stats); - const currJobTickStatus = getAggregateStatus(statusPrecedent, currentJobTick.stats); + // Take groupedBuckets to fill up ticks until we can't anymore + groupedBuckets.forEach((currentGroup, index) => { + const lastTick = jobTicks.at(-1); - // If the current stats are empty and our job tick has reached the min width, finalize the tick - if (statsEmpty && currentJobTick.width >= minTickWidth) { - currentJobTick.roundedRight = true; - jobTicks.push(currentJobTick); - return null; + const lastTickBigEnough = lastTick && lastTick.width >= MINIMUM_TICK_WIDTH; + + const left = index * width; + + const startTs = currentGroup?.at(0)![0]; + const endTs = currentGroup.at(-1)![0] + interval; + const stats = mergeStats(statusPrecedent, ...currentGroup.map(b => b[1])); + + const emptyBucket = isStatsBucketEmpty(stats); + + // Nothing to do if we don't have any data yet + if (emptyBucket && !lastTick) { + return; } - // Calculate the aggregate status for the next minTickWidth buckets - const nextTickAggregateStatus = getAggregateStatusFromMultipleBuckets( - statusPrecedent, - data.slice(i, i + minTickWidth).map(([_, sliceStats]) => sliceStats) - ); - - // If the status changes or we reach the min width, push the current tick and start a new one - if ( - bucketStatus !== currJobTickStatus && - nextTickAggregateStatus !== currJobTickStatus && - currentJobTick.width >= minTickWidth - ) { - jobTicks.push(currentJobTick); - return generateJobTickFromBucketWithStats([timestamp, stats]); + // No data, either expand the previous bucket if it's not big enough or cap + // off the last tick. + if (emptyBucket && lastTick) { + if (lastTickBigEnough === true) { + lastTick.isEnding = true; + } else { + lastTick.endTs = endTs; + lastTick.width += width; + } + return; } - // Otherwise, continue merging data into the current job tick - currentJobTick = { - ...currentJobTick, - endTs: timestamp, - stats: mergeStats(statusPrecedent, currentJobTick.stats, stats), - width: currentJobTick.width + 1, - }; - - // Ensure we render the last tick if it's the final bucket - if (i === data.length - 1) { - currentJobTick.roundedRight = true; - jobTicks.push(currentJobTick); + const startingNewTick = lastTick?.isEnding; + const isFirstTick = !lastTick; + + if (isFirstTick || (startingNewTick && lastTickBigEnough === true)) { + const tick = makeTick({stats, startTs, endTs, left, width, isStarting: true}); + jobTicks.push(tick); + return; + } + + const currentStatus = getAggregateStatus(statusPrecedent, stats); + const lastTickStatus = lastTick?.stats + ? getAggregateStatus(statusPrecedent, lastTick.stats) + : null; + + // We are extending the previous tick if the status's are equal OR if the + // previous bucket has not reached it's minimum size yet. + if (currentStatus === lastTickStatus || lastTickBigEnough === false) { + lastTick.endTs = endTs; + lastTick.stats = mergeStats(statusPrecedent, lastTick.stats, stats); + lastTick.width += width; + + // If we extended the previous tick and the status didn't change there's + // nothing left to do + if (lastTickStatus === getAggregateStatus(statusPrecedent, lastTick.stats)) { + return; + } + + // We've aggregated a new status into the last tick, we may need to merge + // the last tick into other ticks prior to the lastTick, otherwise we may + // end up with multiple ticks that have the same status + while (jobTicks.length > 2) { + const currentTick = jobTicks.at(-1)!; + const priorTick = jobTicks.at(-2)!; + + const currentTickStatus = getAggregateStatus(statusPrecedent, currentTick.stats); + const priorTickStatus = getAggregateStatus(statusPrecedent, priorTick.stats); + + // Nothing to change if the tick prior is an ending tick or has a + // different status from the + if (priorTick.isEnding || currentTickStatus !== priorTickStatus) { + break; + } + + jobTicks.pop()!; + priorTick.endTs = currentTick.endTs; + priorTick.stats = mergeStats(statusPrecedent, priorTick.stats, currentTick.stats); + priorTick.width += currentTick.width; + } + + return; } - return currentJobTick; - }, null); + // Status between the previous tick and the new one is different. Create a + // new tick conjoined to the previous tick. + const tick = makeTick({stats, startTs, endTs, left, width}); + jobTicks.push(tick); + }); return jobTicks; } diff --git a/static/app/components/checkInTimeline/utils/mergeStats.tsx b/static/app/components/checkInTimeline/utils/mergeStats.tsx index 7c80e131a13d29..6b76001ac0b432 100644 --- a/static/app/components/checkInTimeline/utils/mergeStats.tsx +++ b/static/app/components/checkInTimeline/utils/mergeStats.tsx @@ -5,12 +5,14 @@ import type {StatsBucket} from '../types'; */ export function mergeStats<Status extends string>( statusPrecedent: Status[], - statsA: StatsBucket<Status>, - statsB: StatsBucket<Status> + ...stats: Array<StatsBucket<Status>> ): StatsBucket<Status> { const combinedStats = {} as StatsBucket<Status>; for (const status of statusPrecedent) { - combinedStats[status] = (statsA[status] ?? 0) + (statsB[status] ?? 0); + combinedStats[status] = stats.reduce<number>( + (curr, next) => curr + (next[status] ?? 0), + 0 + ); } return combinedStats; } diff --git a/static/app/views/monitors/utils/useMonitorStats.tsx b/static/app/views/monitors/utils/useMonitorStats.tsx index a0710a5a19aa2f..711d3697d13f7b 100644 --- a/static/app/views/monitors/utils/useMonitorStats.tsx +++ b/static/app/views/monitors/utils/useMonitorStats.tsx @@ -20,15 +20,23 @@ interface Options { * Fetches Monitor stats */ export function useMonitorStats({monitors, timeWindowConfig}: Options) { - const {start, end, elapsedMinutes, timelineWidth} = timeWindowConfig; + const {start, end, timelineWidth, rollupConfig} = timeWindowConfig; - // Minimum rollup is 1 second - const rollup = Math.floor((elapsedMinutes * 60) / timelineWidth) || 1; + // Add the underscan to our selection time + const additionalInterval = + (rollupConfig.timelineUnderscanWidth / rollupConfig.bucketPixels) * + rollupConfig.interval; + + // XXX(epurkhiser): We are dropping 1 bucket worth of data on the right side + // to account for the fact that this bucket is actually over-scan becauase + // the query on the backend is inclusive. + const until = + Math.floor(end.getTime() / 1000) + additionalInterval - rollupConfig.interval; const selectionQuery = { since: Math.floor(start.getTime() / 1000), - until: Math.floor(end.getTime() / 1000), - resolution: `${rollup}s`, + until, + resolution: `${rollupConfig.interval}s`, }; const organization = useOrganization(); @@ -42,8 +50,9 @@ export function useMonitorStats({monitors, timeWindowConfig}: Options) { { query: { monitor: monitors, + project: location.query.project, + environment: location.query.environment, ...selectionQuery, - ...location.query, }, }, ],
798c018568abce0b230ecb9e30ed5076d2dc837c
2022-04-13 23:02:22
Kelly Carino
fix(workflow): Handle superuser in team selector on create alert page (#33572)
false
Handle superuser in team selector on create alert page (#33572)
fix
diff --git a/static/app/components/forms/teamSelector.tsx b/static/app/components/forms/teamSelector.tsx index 67c83e9fa8431b..b35f3a8f1058a7 100644 --- a/static/app/components/forms/teamSelector.tsx +++ b/static/app/components/forms/teamSelector.tsx @@ -13,6 +13,7 @@ import {IconAdd, IconUser} from 'sentry/icons'; import {t} from 'sentry/locale'; import space from 'sentry/styles/space'; import {Organization, Project, Team} from 'sentry/types'; +import {isActiveSuperuser} from 'sentry/utils/isActiveSuperuser'; import useApi from 'sentry/utils/useApi'; import useTeams from 'sentry/utils/useTeams'; import withOrganization from 'sentry/utils/withOrganization'; @@ -215,7 +216,12 @@ function TeamSelector(props: Props) { } function getOptions() { - const filteredTeams = teamFilter ? teams.filter(teamFilter) : teams; + const isSuperuser = isActiveSuperuser(); + const filteredTeams = isSuperuser + ? teams + : teamFilter + ? teams.filter(teamFilter) + : teams; if (project) { const teamsInProjectIdSet = new Set(project.teams.map(team => team.id));
67a1a889931ed2faefd87b737f7e8b8b9a6b72eb
2022-04-29 04:33:55
David Wang
feat(widget-builder): Replace GSH with page filters (#33761)
false
Replace GSH with page filters (#33761)
feat
diff --git a/static/app/views/dashboardsV2/widgetBuilder/widgetBuilder.tsx b/static/app/views/dashboardsV2/widgetBuilder/widgetBuilder.tsx index 5cb422568093fa..ff2fd5cb286ac2 100644 --- a/static/app/views/dashboardsV2/widgetBuilder/widgetBuilder.tsx +++ b/static/app/views/dashboardsV2/widgetBuilder/widgetBuilder.tsx @@ -12,11 +12,15 @@ import {addErrorMessage, addSuccessMessage} from 'sentry/actionCreators/indicato import {fetchOrgMembers} from 'sentry/actionCreators/members'; import {loadOrganizationTags} from 'sentry/actionCreators/tags'; import {generateOrderOptions} from 'sentry/components/dashboards/widgetQueriesForm'; +import DatePageFilter from 'sentry/components/datePageFilter'; +import EnvironmentPageFilter from 'sentry/components/environmentPageFilter'; import * as Layout from 'sentry/components/layouts/thirds'; import List from 'sentry/components/list'; import ListItem from 'sentry/components/list/listItem'; import LoadingError from 'sentry/components/loadingError'; +import PageFilterBar from 'sentry/components/organizations/pageFilterBar'; import PageFiltersContainer from 'sentry/components/organizations/pageFilters/container'; +import ProjectPageFilter from 'sentry/components/projectPageFilter'; import SentryDocumentTitle from 'sentry/components/sentryDocumentTitle'; import {t} from 'sentry/locale'; import {PageContent} from 'sentry/styles/organization'; @@ -1003,6 +1007,7 @@ function WidgetBuilder({ defaultSelection={{ datetime: {start: null, end: null, utc: false, period: DEFAULT_STATS_PERIOD}, }} + hideGlobalHeader > <PageContentWithoutPadding> <Header @@ -1017,6 +1022,11 @@ function WidgetBuilder({ <Body> <MainWrapper> <Main> + <StyledPageFilterBar condensed> + <ProjectPageFilter /> + <EnvironmentPageFilter /> + <DatePageFilter alignDropdown="left" /> + </StyledPageFilterBar> <BuildSteps symbol="colored-numeric"> <VisualizationStep widget={currentWidget} @@ -1156,6 +1166,10 @@ const PageContentWithoutPadding = styled(PageContent)` padding: 0; `; +const StyledPageFilterBar = styled(PageFilterBar)` + margin-bottom: ${space(2)}; +`; + const BuildSteps = styled(List)` gap: ${space(4)}; max-width: 100%; diff --git a/tests/js/spec/views/dashboardsV2/widgetBuilder/widgetBuilder.spec.tsx b/tests/js/spec/views/dashboardsV2/widgetBuilder/widgetBuilder.spec.tsx index 600ed32e59d4d1..90ab5eabb563fd 100644 --- a/tests/js/spec/views/dashboardsV2/widgetBuilder/widgetBuilder.spec.tsx +++ b/tests/js/spec/views/dashboardsV2/widgetBuilder/widgetBuilder.spec.tsx @@ -1084,7 +1084,7 @@ describe('WidgetBuilder', function () { expect(handleSave).toHaveBeenCalledTimes(1); }); - it('persists the global selection header period when updating a widget', async () => { + it('persists the page filter period when updating a widget', async () => { const widget: Widget = { id: '1', title: 'Errors over time', @@ -1116,7 +1116,7 @@ describe('WidgetBuilder', function () { }); await screen.findByText('Update Widget'); - await screen.findByText('Last 90 days'); + await screen.findByText('90D'); userEvent.click(screen.getByText('Update Widget'));
6f043bef860b6d0a3870a10af5b384885650db97
2024-03-15 22:20:14
Colleen O'Rourke
ref(rules): Translate int minutes to words (#67021)
false
Translate int minutes to words (#67021)
ref
diff --git a/src/sentry/rules/actions/utils.py b/src/sentry/rules/actions/utils.py index 3013516343e4a8..306772e8f0a2ce 100644 --- a/src/sentry/rules/actions/utils.py +++ b/src/sentry/rules/actions/utils.py @@ -5,6 +5,10 @@ from sentry.models.environment import Environment from sentry.models.rule import Rule +ONE_HOUR = 60 +ONE_DAY = ONE_HOUR * 24 +ONE_WEEK = ONE_DAY * 7 + def get_updated_rule_data(rule: Rule) -> dict[str, Any]: rule_data = dict(rule.data) @@ -42,6 +46,22 @@ def generate_diff_labels( return changed_data +def get_frequency_label(value_str: str | None) -> str | None: + if not value_str: + return None + + value = int(value_str) + if value < 60: + return f"{value} minutes" + elif value >= 60 and value < 10080: + return f"{int(value / ONE_HOUR)} hours" + elif value == ONE_WEEK: + return f"{int(value / ONE_WEEK)} week" + elif value == ONE_DAY * 30: + return f"{int(value / ONE_DAY)} days" + return None + + def get_changed_data( rule: Rule, rule_data: dict[str, Any], rule_data_before: dict[str, Any] ) -> dict[str, Any]: @@ -62,8 +82,10 @@ def get_changed_data( rule_data, rule_data_before, rule, changed_data, "actions", "Removed action '{}'" ) - frequency_text = check_value_changed(rule_data, rule_data_before, "frequency", "frequency") - if frequency_text: + current_frequency = get_frequency_label(rule_data.get("frequency")) + previous_frequency = get_frequency_label(rule_data_before.get("frequency")) + if current_frequency != previous_frequency: + frequency_text = f"Changed frequency from *{previous_frequency}* to *{current_frequency}*" changed_data["changed_frequency"].append(frequency_text) if rule_data.get("environment_id") and not rule_data_before.get("environment_id"): diff --git a/tests/sentry/api/endpoints/test_project_rule_details.py b/tests/sentry/api/endpoints/test_project_rule_details.py index 5af1f58a9e402a..2a35a293281b49 100644 --- a/tests/sentry/api/endpoints/test_project_rule_details.py +++ b/tests/sentry/api/endpoints/test_project_rule_details.py @@ -1013,7 +1013,9 @@ def test_slack_confirmation_notification_contents(self): "channel": "#old_channel_name", } ] - self.rule.update(data={"conditions": conditions, "actions": actions}, label="my rule") + self.rule.update( + data={"conditions": conditions, "actions": actions, "frequency": 5}, label="my rule" + ) actions[0]["channel"] = "#new_channel_name" actions[0]["channel_id"] = "new_channel_id" @@ -1063,7 +1065,7 @@ def test_slack_confirmation_notification_contents(self): "filterMatch": "any", "actions": actions, "conditions": conditions, - "frequency": 30, + "frequency": 180, "environment": staging_env.name, "owner": get_actor_for_user(self.user).get_actor_identifier(), } @@ -1081,7 +1083,7 @@ def test_slack_confirmation_notification_contents(self): changes = "*Changes*\n" changes += "• Added action 'Send a notification to the Awesome Team Slack workspace to new_channel_name (optionally, an ID: new_channel_id) and show tags [] in notification'\n" changes += "• Removed action 'Send a notification to the Awesome Team Slack workspace to #old_channel_name (optionally, an ID: old_channel_id) and show tags [] in notification'\n" - changes += "• Changed frequency from *None* to *30*\n" + changes += "• Changed frequency from *5 minutes* to *3 hours*\n" changes += f"• Added *{staging_env.name}* environment\n" changes += "• Changed rule name from *my rule* to *new rule*\n" changes += "• Changed trigger from *None* to *any*\n"
65823cb7837f54df3969af518363e38741c5d2cf
2018-09-25 00:14:51
Lyn Nagara
feat(discover): Add endpoint to get a saved query (#9831)
false
Add endpoint to get a saved query (#9831)
feat
diff --git a/src/sentry/api/endpoints/organization_discover_saved_query_detail.py b/src/sentry/api/endpoints/organization_discover_saved_query_detail.py new file mode 100644 index 00000000000000..e6abeaa8898ce6 --- /dev/null +++ b/src/sentry/api/endpoints/organization_discover_saved_query_detail.py @@ -0,0 +1,27 @@ +from __future__ import absolute_import + +from rest_framework.response import Response +from sentry.api.serializers import serialize +from sentry.api.exceptions import ResourceDoesNotExist +from sentry.api.bases.organization import OrganizationPermission +from sentry.api.bases import OrganizationEndpoint +from sentry import features +from sentry.models import DiscoverSavedQuery + + +class OrganizationDiscoverSavedQueryDetailEndpoint(OrganizationEndpoint): + permission_classes = (OrganizationPermission, ) + + def get(self, request, organization, query_id): + """ + Get a saved query + """ + if not features.has('organizations:discover', organization, actor=request.user): + return self.respond(status=404) + + try: + item = DiscoverSavedQuery.objects.get(id=query_id, organization=organization) + except DiscoverSavedQuery.DoesNotExist: + raise ResourceDoesNotExist + + return Response(serialize(item), status=200) diff --git a/src/sentry/api/urls.py b/src/sentry/api/urls.py index 09186713fd1230..ee9e6547e20e48 100644 --- a/src/sentry/api/urls.py +++ b/src/sentry/api/urls.py @@ -60,6 +60,7 @@ from .endpoints.organization_details import OrganizationDetailsEndpoint from .endpoints.organization_discover_query import OrganizationDiscoverQueryEndpoint from .endpoints.organization_discover_saved_queries import OrganizationDiscoverSavedQueriesEndpoint +from .endpoints.organization_discover_saved_query_detail import OrganizationDiscoverSavedQueryDetailEndpoint from .endpoints.organization_health import OrganizationHealthTopEndpoint, OrganizationHealthGraphEndpoint from .endpoints.organization_shortid import ShortIdLookupEndpoint from .endpoints.organization_environments import OrganizationEnvironmentsEndpoint @@ -385,6 +386,11 @@ OrganizationDiscoverSavedQueriesEndpoint.as_view(), name='sentry-api-0-organization-discover-saved-queries' ), + url( + r'^organizations/(?P<organization_slug>[^\/]+)/discover/saved/(?P<query_id>[^\/]+)/$', + OrganizationDiscoverSavedQueryDetailEndpoint.as_view(), + name='sentry-api-0-organization-discover-saved-query-detail' + ), url( r'^organizations/(?P<organization_slug>[^\/]+)/health/top/$', OrganizationHealthTopEndpoint.as_view(), diff --git a/tests/snuba/test_organization_discover_saved_query_detail.py b/tests/snuba/test_organization_discover_saved_query_detail.py new file mode 100644 index 00000000000000..2b1e5db6449b50 --- /dev/null +++ b/tests/snuba/test_organization_discover_saved_query_detail.py @@ -0,0 +1,59 @@ +from __future__ import absolute_import + +import six +from sentry.testutils import APITestCase +from django.core.urlresolvers import reverse +from sentry.testutils import SnubaTestCase + +from sentry.models import DiscoverSavedQuery + + +class OrganizationDiscoverSavedQueryDetailTest(APITestCase, SnubaTestCase): + def setUp(self): + super(OrganizationDiscoverSavedQueryDetailTest, self).setUp() + self.login_as(user=self.user) + self.org = self.create_organization(owner=self.user) + self.org_without_access = self.create_organization() + self.project_ids = [ + self.create_project(organization=self.org).id, + self.create_project(organization=self.org).id + ] + query = { + 'fields': ['test'], + 'conditions': [], + 'limit': 10 + } + + model = DiscoverSavedQuery.objects.create( + organization=self.org, name="Test query", query=query) + + model.add_projects(self.project_ids) + + self.query_id = model.id + + def test_get(self): + with self.feature('organizations:discover'): + url = reverse( + 'sentry-api-0-organization-discover-saved-query-detail', + args=[ + self.org.slug, + self.query_id]) + response = self.client.get(url) + + assert response.status_code == 200, response.content + assert response.data['id'] == six.text_type(self.query_id) + assert response.data['projects'] == self.project_ids + assert response.data['fields'] == ['test'] + assert response.data['conditions'] == [] + assert response.data['limit'] == 10 + + def test_get_org_without_access(self): + with self.feature('organizations:discover'): + url = reverse( + 'sentry-api-0-organization-discover-saved-query-detail', + args=[ + self.org_without_access.slug, + self.query_id]) + response = self.client.get(url) + + assert response.status_code == 403, response.content
6cf39fb26b15d4c8161117be8eb4bf10a0985304
2017-09-20 13:08:23
Matt Robenolt
fix(devserver): Don't attempt to bind `https` to ports it's not allowed
false
Don't attempt to bind `https` to ports it's not allowed
fix
diff --git a/src/sentry/runner/commands/devserver.py b/src/sentry/runner/commands/devserver.py index f36ace556804c7..f543b459a5bb63 100644 --- a/src/sentry/runner/commands/devserver.py +++ b/src/sentry/runner/commands/devserver.py @@ -9,6 +9,7 @@ import click import six +from six.moves.urllib.parse import urlparse from sentry.runner.decorators import configuration, log_options @@ -51,7 +52,12 @@ def devserver(reload, watchers, workers, browser_reload, styleguide, environment from sentry.services.http import SentryHTTPServer url_prefix = options.get('system.url-prefix', '') - needs_https = url_prefix.startswith('https://') + parsed_url = urlparse(url_prefix) + # Make sure we're trying to use a port that we can actually bind to + needs_https = ( + parsed_url.scheme == 'https' and + (parsed_url.port or 443) > 1024 + ) has_https = False if needs_https: @@ -99,9 +105,7 @@ def devserver(reload, watchers, workers, browser_reload, styleguide, environment ] if needs_https and has_https: - from six.moves.urllib.parse import urlparse - parsed_url = urlparse(url_prefix) - https_port = six.text_type(parsed_url.port or 443) + https_port = six.text_type(parsed_url.port) https_host = parsed_url.hostname # Determine a random port for the backend http server
034215a635b263af20fd11a512b3bc223bc0dd74
2024-10-19 01:55:25
Evan Purkhiser
ref(rr6): Remove unused param (#79379)
false
Remove unused param (#79379)
ref
diff --git a/static/app/bootstrap/initializeSdk.tsx b/static/app/bootstrap/initializeSdk.tsx index 9ddba27c967414..84cfa3f7d95b40 100644 --- a/static/app/bootstrap/initializeSdk.tsx +++ b/static/app/bootstrap/initializeSdk.tsx @@ -78,15 +78,13 @@ function getSentryIntegrations() { return integrations; } -// TODO(__SENTRY_USING_REACT_ROUTER_SIX): Remove opts once getsentry has had -// this paramter removed /** * Initialize the Sentry SDK * * If `routes` is passed, we will instrument react-router. Not all * entrypoints require this. */ -export function initializeSdk(config: Config, _otps?: any) { +export function initializeSdk(config: Config) { const {apmSampling, sentryConfig, userIdentity} = config; const tracesSampleRate = apmSampling ?? 0; const extraTracePropagationTargets = SPA_DSN
dac1aa477e46153465498469c6ad9088770464ba
2024-02-15 00:13:48
Shruthi
chore(spans): Tag nodestore requests to track usage (#65136)
false
Tag nodestore requests to track usage (#65136)
chore
diff --git a/src/sentry/eventstore/base.py b/src/sentry/eventstore/base.py index 9fb27bb62cc746..90200db81fafa5 100644 --- a/src/sentry/eventstore/base.py +++ b/src/sentry/eventstore/base.py @@ -267,6 +267,8 @@ def bind_nodes(self, object_list, node_name="data"): It's not necessary to bind a single Event object since data will be lazily fetched on any attempt to access a property. """ + sentry_sdk.set_tag("eventstore.backend", "nodestore") + with sentry_sdk.start_span(op="eventstore.base.bind_nodes"): object_node_list = [ (i, getattr(i, node_name)) for i in object_list if getattr(i, node_name).id
70cabebde7a5e922bec2b1840025e256643393f4
2022-12-21 05:04:56
Elias Hussary
fix(profiling): guard against selecting sentry root with keyboard nav (#42495)
false
guard against selecting sentry root with keyboard nav (#42495)
fix
diff --git a/static/app/components/profiling/flamegraph/flamegraphZoomView.tsx b/static/app/components/profiling/flamegraph/flamegraphZoomView.tsx index f42d80278e18bb..ba829190945e59 100644 --- a/static/app/components/profiling/flamegraph/flamegraphZoomView.tsx +++ b/static/app/components/profiling/flamegraph/flamegraphZoomView.tsx @@ -870,7 +870,7 @@ function handleFlamegraphKeyboardNavigation( direction = direction === 'up' ? 'down' : 'up'; } const nextSelection = selectNearestFrame(currentFrame, direction); - if (!nextSelection) { + if (nextSelection === currentFrame) { return null; } evt.preventDefault(); diff --git a/static/app/utils/profiling/flamegraph/selectNearestFrame.spec.ts b/static/app/utils/profiling/flamegraph/selectNearestFrame.spec.ts index 496c8d38cc3c62..a7d3b900f7f090 100644 --- a/static/app/utils/profiling/flamegraph/selectNearestFrame.spec.ts +++ b/static/app/utils/profiling/flamegraph/selectNearestFrame.spec.ts @@ -1,10 +1,22 @@ +import {DeepPartial} from 'sentry/types/utils'; + import {FlamegraphFrame} from '../flamegraphFrame'; import {selectNearestFrame} from './selectNearestFrame'; -function createFlamegraphFrame(frame?: Partial<FlamegraphFrame>) { - const {depth = 0, parent = null, children = []} = frame ?? {}; +function createFlamegraphFrame(frame?: DeepPartial<FlamegraphFrame>) { + const { + depth = 0, + parent = null, + children = [], + frame: _frame = { + isRoot() { + return false; + }, + }, + } = frame ?? {}; return { + frame: _frame, depth, parent, children, @@ -101,4 +113,17 @@ describe('selectNearestFrame', () => { const next = selectNearestFrame(leftGrandChild as any, 'down'); expect(next).toBe(rightGrandChild.parent); }); + + it('does not allow selection of the "sentry root" virtual root node', () => { + const root = createFlamegraphFrame({ + frame: { + isRoot() { + return true; + }, + }, + }); + const leftChild = addChildrenToDepth(root, 1); + const next = selectNearestFrame(leftChild as any, 'up'); + expect(next).toBe(leftChild); + }); }); diff --git a/static/app/utils/profiling/flamegraph/selectNearestFrame.ts b/static/app/utils/profiling/flamegraph/selectNearestFrame.ts index 8d958682eaab70..16e5f19ddb1d84 100644 --- a/static/app/utils/profiling/flamegraph/selectNearestFrame.ts +++ b/static/app/utils/profiling/flamegraph/selectNearestFrame.ts @@ -10,6 +10,12 @@ export type Direction = DirectionY | DirectionX; export function selectNearestFrame(frame: FlamegraphFrame, direction: Direction) { if (direction === 'up') { const parent = frame.parent; + + // sentry root is a virtual root that should not be selectable + if (parent?.frame.isRoot()) { + return frame; + } + // if there is an immediate parent, goto parent if (parent) { return parent;
5d00d38a54a341e826d59205e2a40aff4a7022ae
2022-10-12 20:09:18
Stephen Cefali
feat(code-owners): remove request to set up codeowners (#39897)
false
remove request to set up codeowners (#39897)
feat
diff --git a/static/app/views/settings/project/projectOwnership/index.spec.jsx b/static/app/views/settings/project/projectOwnership/index.spec.jsx index 6d6ddf5665bdd5..e1a81cb2a0804a 100644 --- a/static/app/views/settings/project/projectOwnership/index.spec.jsx +++ b/static/app/views/settings/project/projectOwnership/index.spec.jsx @@ -91,7 +91,7 @@ describe('Project Ownership', function () { expect(openModal).toHaveBeenCalled(); }); - it('render request to add if no permissions', function () { + it('render nothing to add if no permissions', function () { org = TestStubs.Organization({features: ['integrations-codeowners'], access: []}); const wrapper = mountWithTheme( @@ -102,9 +102,9 @@ describe('Project Ownership', function () { />, TestStubs.routerContext([{organization: org}]) ); - expect( - wrapper.find('[data-test-id="add-codeowner-request-button"] button').exists() - ).toBe(true); + expect(wrapper.find('[data-test-id="add-codeowner-button"] button').exists()).toBe( + false + ); }); }); }); diff --git a/static/app/views/settings/project/projectOwnership/index.tsx b/static/app/views/settings/project/projectOwnership/index.tsx index d8f2dd32f9ff76..9e464383f3eaff 100644 --- a/static/app/views/settings/project/projectOwnership/index.tsx +++ b/static/app/views/settings/project/projectOwnership/index.tsx @@ -1,13 +1,7 @@ import {Fragment} from 'react'; import {RouteComponentProps} from 'react-router'; import styled from '@emotion/styled'; -import * as Sentry from '@sentry/react'; -import { - addErrorMessage, - addLoadingMessage, - addSuccessMessage, -} from 'sentry/actionCreators/indicator'; import {openEditOwnershipRules, openModal} from 'sentry/actionCreators/modal'; import Access from 'sentry/components/acl/access'; import Feature from 'sentry/components/acl/feature'; @@ -123,25 +117,6 @@ tags.sku_class:enterprise #enterprise`; }); }; - handleAddCodeOwnerRequest = async () => { - const {organization, project} = this.props; - try { - addLoadingMessage(t('Requesting\u2026')); - await this.api.requestPromise( - `/projects/${organization.slug}/${project.slug}/codeowners-request/`, - { - method: 'POST', - data: {}, - } - ); - - addSuccessMessage(t('Request Sent')); - } catch (err) { - addErrorMessage(t('Unable to send request')); - Sentry.captureException(err); - } - }; - renderCodeOwnerErrors = () => { const {project, organization} = this.props; const {codeowners} = this.state; @@ -291,16 +266,7 @@ tags.sku_class:enterprise #enterprise`; > {t('Add CODEOWNERS')} </CodeOwnerButton> - ) : ( - <CodeOwnerButton - onClick={this.handleAddCodeOwnerRequest} - size="sm" - priority="primary" - data-test-id="add-codeowner-request-button" - > - {t('Request to Add CODEOWNERS File')} - </CodeOwnerButton> - ) + ) : null } </Access> </Feature>
792981b2148eb3256109f683b4d8ef931d460616
2025-01-16 04:13:53
Josh Callender
feat(workflow_engine): Add a `DataConditionHandlerType` to each DataConditionHandler (#83276)
false
Add a `DataConditionHandlerType` to each DataConditionHandler (#83276)
feat
diff --git a/src/sentry/workflow_engine/handlers/condition/age_comparison_handler.py b/src/sentry/workflow_engine/handlers/condition/age_comparison_handler.py index 6c50e125793137..5f9adf00d6ed2c 100644 --- a/src/sentry/workflow_engine/handlers/condition/age_comparison_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/age_comparison_handler.py @@ -6,11 +6,13 @@ from sentry.rules.filters.age_comparison import timeranges from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.AGE_COMPARISON) class AgeComparisonConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + comparison_json_schema = { "type": "object", "properties": { diff --git a/src/sentry/workflow_engine/handlers/condition/assigned_to_handler.py b/src/sentry/workflow_engine/handlers/condition/assigned_to_handler.py index 4b39fe5d6e2f04..e406dd235bec1f 100644 --- a/src/sentry/workflow_engine/handlers/condition/assigned_to_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/assigned_to_handler.py @@ -7,11 +7,13 @@ from sentry.utils.cache import cache from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.ASSIGNED_TO) class AssignedToConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + @staticmethod def get_assignees(group: Group) -> Sequence[GroupAssignee]: cache_key = f"group:{group.id}:assignees" diff --git a/src/sentry/workflow_engine/handlers/condition/event_attribute_handler.py b/src/sentry/workflow_engine/handlers/condition/event_attribute_handler.py index e997537bfd2c38..51b986e5b80b8d 100644 --- a/src/sentry/workflow_engine/handlers/condition/event_attribute_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/event_attribute_handler.py @@ -8,11 +8,13 @@ from sentry.utils.registry import NoRegistrationExistsError from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.EVENT_ATTRIBUTE) class EventAttributeConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + @staticmethod def get_attribute_values(event: GroupEvent, attribute: str) -> list[str]: path = attribute.split(".") diff --git a/src/sentry/workflow_engine/handlers/condition/event_created_by_detector_handler.py b/src/sentry/workflow_engine/handlers/condition/event_created_by_detector_handler.py index 2334293b2b379a..0f5defdc1e4db4 100644 --- a/src/sentry/workflow_engine/handlers/condition/event_created_by_detector_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/event_created_by_detector_handler.py @@ -2,11 +2,13 @@ from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.EVENT_CREATED_BY_DETECTOR) class EventCreatedByDetectorConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: event = job["event"] diff --git a/src/sentry/workflow_engine/handlers/condition/event_seen_count_handler.py b/src/sentry/workflow_engine/handlers/condition/event_seen_count_handler.py index 1e426e80476d75..d328ef59f85773 100644 --- a/src/sentry/workflow_engine/handlers/condition/event_seen_count_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/event_seen_count_handler.py @@ -2,11 +2,13 @@ from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.EVENT_SEEN_COUNT) class EventSeenCountConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: event = job["event"] diff --git a/src/sentry/workflow_engine/handlers/condition/every_event_handler.py b/src/sentry/workflow_engine/handlers/condition/every_event_handler.py index 62df43141cb998..969c684bd7305f 100644 --- a/src/sentry/workflow_engine/handlers/condition/every_event_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/every_event_handler.py @@ -2,11 +2,13 @@ from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.EVERY_EVENT) class EveryEventConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.WORKFLOW_TRIGGER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: return True diff --git a/src/sentry/workflow_engine/handlers/condition/existing_high_priority_issue_handler.py b/src/sentry/workflow_engine/handlers/condition/existing_high_priority_issue_handler.py index 06dd226f69a0f7..4b45c72a8861a1 100644 --- a/src/sentry/workflow_engine/handlers/condition/existing_high_priority_issue_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/existing_high_priority_issue_handler.py @@ -3,11 +3,13 @@ from sentry.types.group import PriorityLevel from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.EXISTING_HIGH_PRIORITY_ISSUE) class ExistingHighPriorityIssueConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.WORKFLOW_TRIGGER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: state = job.get("group_state") diff --git a/src/sentry/workflow_engine/handlers/condition/first_seen_event_handler.py b/src/sentry/workflow_engine/handlers/condition/first_seen_event_handler.py index ebc2f154e405c5..1261d3cbef854b 100644 --- a/src/sentry/workflow_engine/handlers/condition/first_seen_event_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/first_seen_event_handler.py @@ -2,7 +2,7 @@ from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob def is_new_event(job: WorkflowJob) -> bool: @@ -19,6 +19,8 @@ def is_new_event(job: WorkflowJob) -> bool: @condition_handler_registry.register(Condition.FIRST_SEEN_EVENT) class FirstSeenEventConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.WORKFLOW_TRIGGER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: return is_new_event(job) diff --git a/src/sentry/workflow_engine/handlers/condition/issue_occurrences_handler.py b/src/sentry/workflow_engine/handlers/condition/issue_occurrences_handler.py index d4d43621844519..fcab4a0cb887ce 100644 --- a/src/sentry/workflow_engine/handlers/condition/issue_occurrences_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/issue_occurrences_handler.py @@ -3,11 +3,13 @@ from sentry.models.group import Group from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.ISSUE_OCCURRENCES) class IssueOccurrencesConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: group: Group = job["event"].group diff --git a/src/sentry/workflow_engine/handlers/condition/latest_release_handler.py b/src/sentry/workflow_engine/handlers/condition/latest_release_handler.py index 5e7ee50ea0f8b8..61548897ab198e 100644 --- a/src/sentry/workflow_engine/handlers/condition/latest_release_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/latest_release_handler.py @@ -9,7 +9,7 @@ from sentry.utils.cache import cache from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob def get_latest_release_for_env( @@ -41,6 +41,8 @@ def get_latest_release_for_env( @condition_handler_registry.register(Condition.LATEST_RELEASE) class LatestReleaseConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: event = job["event"] diff --git a/src/sentry/workflow_engine/handlers/condition/level_handler.py b/src/sentry/workflow_engine/handlers/condition/level_handler.py index 208860e00ab5b8..af77805c03e546 100644 --- a/src/sentry/workflow_engine/handlers/condition/level_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/level_handler.py @@ -4,11 +4,13 @@ from sentry.rules import MatchType from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.LEVEL) class LevelConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: event = job["event"] diff --git a/src/sentry/workflow_engine/handlers/condition/new_high_priority_issue_handler.py b/src/sentry/workflow_engine/handlers/condition/new_high_priority_issue_handler.py index dd981b976cfbdf..9a2b902e73dc9c 100644 --- a/src/sentry/workflow_engine/handlers/condition/new_high_priority_issue_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/new_high_priority_issue_handler.py @@ -4,11 +4,13 @@ from sentry.workflow_engine.handlers.condition.first_seen_event_handler import is_new_event from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.NEW_HIGH_PRIORITY_ISSUE) class NewHighPriorityIssueConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.WORKFLOW_TRIGGER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: is_new = is_new_event(job) diff --git a/src/sentry/workflow_engine/handlers/condition/reappeared_event_handler.py b/src/sentry/workflow_engine/handlers/condition/reappeared_event_handler.py index 1c52be0947b577..6c4be4211e50d8 100644 --- a/src/sentry/workflow_engine/handlers/condition/reappeared_event_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/reappeared_event_handler.py @@ -2,11 +2,13 @@ from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.REAPPEARED_EVENT) class ReappearedEventConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.WORKFLOW_TRIGGER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: has_reappeared = job.get("has_reappeared") diff --git a/src/sentry/workflow_engine/handlers/condition/regression_event_handler.py b/src/sentry/workflow_engine/handlers/condition/regression_event_handler.py index d09cea50c944af..603bec0e147c76 100644 --- a/src/sentry/workflow_engine/handlers/condition/regression_event_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/regression_event_handler.py @@ -2,11 +2,13 @@ from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.REGRESSION_EVENT) class RegressionEventConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.WORKFLOW_TRIGGER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: state = job.get("group_state") diff --git a/src/sentry/workflow_engine/handlers/condition/tagged_event_handler.py b/src/sentry/workflow_engine/handlers/condition/tagged_event_handler.py index 803f7bb83891e7..2a5a34ee28dd9d 100644 --- a/src/sentry/workflow_engine/handlers/condition/tagged_event_handler.py +++ b/src/sentry/workflow_engine/handlers/condition/tagged_event_handler.py @@ -4,11 +4,13 @@ from sentry.rules import MatchType, match_values from sentry.workflow_engine.models.data_condition import Condition from sentry.workflow_engine.registry import condition_handler_registry -from sentry.workflow_engine.types import DataConditionHandler, WorkflowJob +from sentry.workflow_engine.types import DataConditionHandler, DataConditionHandlerType, WorkflowJob @condition_handler_registry.register(Condition.TAGGED_EVENT) class TaggedEventConditionHandler(DataConditionHandler[WorkflowJob]): + type = DataConditionHandlerType.ACTION_FILTER + @staticmethod def evaluate_value(job: WorkflowJob, comparison: Any) -> bool: event = job["event"] diff --git a/src/sentry/workflow_engine/types.py b/src/sentry/workflow_engine/types.py index a5833a47916bc7..c569455d5815be 100644 --- a/src/sentry/workflow_engine/types.py +++ b/src/sentry/workflow_engine/types.py @@ -1,6 +1,6 @@ from __future__ import annotations -from enum import IntEnum +from enum import IntEnum, StrEnum from typing import TYPE_CHECKING, Any, ClassVar, Generic, TypedDict, TypeVar from sentry.types.group import PriorityLevel @@ -20,6 +20,12 @@ class DetectorPriorityLevel(IntEnum): HIGH = PriorityLevel.HIGH +class DataConditionHandlerType(StrEnum): + DETECTOR_TRIGGER = "detector_trigger" + WORKFLOW_TRIGGER = "workflow_trigger" + ACTION_FILTER = "action_filter" + + # The unique key used to identify a group within a DataPacket result. # For DataPackets that don't contain multiple values the key is just None. # This is stored in 'DetectorState.detector_group_key' @@ -55,6 +61,7 @@ def bulk_get_query_object(data_sources) -> dict[int, T | None]: class DataConditionHandler(Generic[T]): + type: ClassVar[DataConditionHandlerType] = DataConditionHandlerType.ACTION_FILTER comparison_json_schema: ClassVar[dict[str, Any]] = {} @staticmethod
6b3efbfd0f61d54e5fbd3d8d4b4fc2ff6338d1cd
2021-05-27 19:51:55
Joris Bayer
feat(releases): Store archives as release files (#25849)
false
Store archives as release files (#25849)
feat
diff --git a/src/sentry/conf/server.py b/src/sentry/conf/server.py index 73041267c6d287..23b940f9410f6c 100644 --- a/src/sentry/conf/server.py +++ b/src/sentry/conf/server.py @@ -996,6 +996,8 @@ def create_partitioned_queues(name): "organizations:alert-wizard": True, # Enable the adoption chart in the releases page "organizations:release-adoption-chart": False, + # Store release bundles as zip files instead of single files + "organizations:release-archives": False, # Enable the project level transaction thresholds "organizations:project-transaction-threshold": False, # Enable percent displays in issue stream diff --git a/src/sentry/features/__init__.py b/src/sentry/features/__init__.py index a38e000ca298d0..73793adeee02dc 100644 --- a/src/sentry/features/__init__.py +++ b/src/sentry/features/__init__.py @@ -110,6 +110,7 @@ default_manager.add("organizations:related-events", OrganizationFeature) # NOQA default_manager.add("organizations:release-adoption-chart", OrganizationFeature, True) # NOQA default_manager.add("organizations:relay", OrganizationFeature) # NOQA +default_manager.add("organizations:release-archives", OrganizationFeature) # NOQA default_manager.add("organizations:reprocessing-v2", OrganizationFeature) # NOQA default_manager.add("organizations:rule-page", OrganizationFeature) # NOQA default_manager.add("organizations:set-grouping-config", OrganizationFeature) # NOQA diff --git a/src/sentry/lang/javascript/processor.py b/src/sentry/lang/javascript/processor.py index 4873e7015016c0..a3a6f67737130a 100644 --- a/src/sentry/lang/javascript/processor.py +++ b/src/sentry/lang/javascript/processor.py @@ -1,5 +1,13 @@ +import random +import time +from io import BytesIO +from typing import IO, Optional, Tuple + from django.utils.encoding import force_bytes, force_text +from sentry.models.releasefile import ReleaseArchive +from sentry.tasks.assemble import RELEASE_ARCHIVE_FILENAME + __all__ = ["JavaScriptStacktraceProcessor"] import base64 @@ -16,7 +24,7 @@ from requests.utils import get_encoding_from_headers from symbolic import SourceMapView -from sentry import http +from sentry import http, options from sentry.interfaces.stacktrace import Stacktrace from sentry.models import EventError, Organization, ReleaseFile from sentry.stacktraces.processing import StacktraceProcessor @@ -221,18 +229,58 @@ def should_retry_fetch(attempt: int, e: Exception) -> bool: fetch_retry_policy = ConditionalRetryPolicy(should_retry_fetch, exponential_delay(0.05)) -def fetch_release_file(filename, release, dist=None): - """ - Attempt to retrieve a release artifact from the database. +def fetch_and_cache_artifact(filename, fetch_fn, cache_key, cache_key_meta, headers, compress_fn): + # If the release file is not in cache, check if we can retrieve at + # least the size metadata from cache and prevent compression and + # caching if payload exceeds the backend limit. + z_body_size = None + + if CACHE_MAX_VALUE_SIZE: + cache_meta = cache.get(cache_key_meta) + if cache_meta: + z_body_size = int(cache_meta.get("compressed_size")) + + def fetch_release_body(): + with fetch_fn() as fp: + if z_body_size and z_body_size > CACHE_MAX_VALUE_SIZE: + return None, fp.read() + else: + return compress_fn(fp) + + try: + with metrics.timer("sourcemaps.release_file_read"): + z_body, body = fetch_retry_policy(fetch_release_body) + except Exception: + logger.error("sourcemap.compress_read_failed", exc_info=sys.exc_info()) + result = None + else: + headers = {k.lower(): v for k, v in headers.items()} + encoding = get_encoding_from_headers(headers) + result = http.UrlResult(filename, headers, body, 200, encoding) + + # If we don't have the compressed body for caching because the + # cached metadata said it is too large payload for the cache + # backend, do not attempt to cache. + if z_body: + # This will implicitly skip too large payloads. Those will be cached + # on the file system by `ReleaseFile.cache`, instead. + cache.set(cache_key, (headers, z_body, 200, encoding), 3600) + + # In case the previous call to cache implicitly fails, we use + # the meta data to avoid pointless compression which is done + # only for caching. + cache.set(cache_key_meta, {"compressed_size": len(z_body)}, 3600) + + return result - Caches the result of that attempt (whether successful or not). - """ +def get_cache_keys(filename, release, dist): dist_name = dist and dist.name or None releasefile_ident = ReleaseFile.get_ident(filename, dist_name) cache_key = get_release_file_cache_key( release_id=release.id, releasefile_ident=releasefile_ident ) + # Cache key to store file metadata, currently only the size of the # compressed version of file. We cannot use the cache_key because large # payloads (silently) fail to cache due to e.g. memcached payload size @@ -241,86 +289,84 @@ def fetch_release_file(filename, release, dist=None): release_id=release.id, releasefile_ident=releasefile_ident ) + return cache_key, cache_key_meta + + +def result_from_cache(filename, result): + # Previous caches would be a 3-tuple instead of a 4-tuple, + # so this is being maintained for backwards compatibility + try: + encoding = result[3] + except IndexError: + encoding = None + + return http.UrlResult(filename, result[0], zlib.decompress(result[1]), result[2], encoding) + + [email protected]("sourcemaps.release_file") +def fetch_release_file(filename, release, dist=None): + """ + Attempt to retrieve a release artifact from the database. + + Caches the result of that attempt (whether successful or not). + """ + dist_name = dist and dist.name or None + cache_key, cache_key_meta = get_cache_keys(filename, release, dist) + logger.debug("Checking cache for release artifact %r (release_id=%s)", filename, release.id) result = cache.get(cache_key) # not in the cache (meaning we haven't checked the database recently), so check the database if result is None: - filename_choices = ReleaseFile.normalize(filename) - filename_idents = [ReleaseFile.get_ident(f, dist_name) for f in filename_choices] - - logger.debug( - "Checking database for release artifact %r (release_id=%s)", filename, release.id - ) + with metrics.timer("sourcemaps.release_artifact_from_file"): + filename_choices = ReleaseFile.normalize(filename) + filename_idents = [ReleaseFile.get_ident(f, dist_name) for f in filename_choices] - possible_files = list( - ReleaseFile.objects.filter( - release=release, dist=dist, ident__in=filename_idents - ).select_related("file") - ) - - if len(possible_files) == 0: logger.debug( - "Release artifact %r not found in database (release_id=%s)", filename, release.id + "Checking database for release artifact %r (release_id=%s)", filename, release.id ) - cache.set(cache_key, -1, 60) - return None - elif len(possible_files) == 1: - releasefile = possible_files[0] - - else: - # Pick first one that matches in priority order. - # This is O(N*M) but there are only ever at most 4 things here - # so not really worth optimizing. - releasefile = next( - rf for ident in filename_idents for rf in possible_files if rf.ident == ident + possible_files = list( + ReleaseFile.objects.filter( + release=release, dist=dist, ident__in=filename_idents + ).select_related("file") ) - logger.debug( - "Found release artifact %r (id=%s, release_id=%s)", filename, releasefile.id, release.id - ) + if len(possible_files) == 0: + logger.debug( + "Release artifact %r not found in database (release_id=%s)", + filename, + release.id, + ) + cache.set(cache_key, -1, 60) + return None - # If the release file is not in cache, check if we can retrieve at - # least the size metadata from cache and prevent compression and - # caching if payload exceeds the backend limit. - z_body_size = None + elif len(possible_files) == 1: + releasefile = possible_files[0] - if CACHE_MAX_VALUE_SIZE: - cache_meta = cache.get(cache_key_meta) - if cache_meta: - z_body_size = int(cache_meta.get("compressed_size")) + else: + # Pick first one that matches in priority order. + # This is O(N*M) but there are only ever at most 4 things here + # so not really worth optimizing. + releasefile = next( + rf for ident in filename_idents for rf in possible_files if rf.ident == ident + ) - def fetch_release_body(): - with ReleaseFile.cache.getfile(releasefile) as fp: - if z_body_size and z_body_size > CACHE_MAX_VALUE_SIZE: - return None, fp.read() - else: - return compress_file(fp) + logger.debug( + "Found release artifact %r (id=%s, release_id=%s)", + filename, + releasefile.id, + release.id, + ) - try: - with metrics.timer("sourcemaps.release_file_read"): - z_body, body = fetch_retry_policy(fetch_release_body) - except Exception: - logger.error("sourcemap.compress_read_failed", exc_info=sys.exc_info()) - result = None - else: - headers = {k.lower(): v for k, v in releasefile.file.headers.items()} - encoding = get_encoding_from_headers(headers) - result = http.UrlResult(filename, headers, body, 200, encoding) - - # If we don't have the compressed body for caching because the - # cached metadata said it is too large payload for the cache - # backend, do not attempt to cache. - if z_body: - # This will implicitly skip too large payloads. Those will be cached - # on the file system by `ReleaseFile.cache`, instead. - cache.set(cache_key, (headers, z_body, 200, encoding), 3600) - - # In case the previous call to cache implicitly fails, we use - # the meta data to avoid pointless compression which is done - # only for caching. - cache.set(cache_key_meta, {"compressed_size": len(z_body)}, 3600) + result = fetch_and_cache_artifact( + filename, + lambda: ReleaseFile.cache.getfile(releasefile), + cache_key, + cache_key_meta, + releasefile.file.headers, + compress_file, + ) # in the cache as an unsuccessful attempt elif result == -1: @@ -328,15 +374,135 @@ def fetch_release_body(): # in the cache as a successful attempt, including the zipped contents of the file else: - # Previous caches would be a 3-tuple instead of a 4-tuple, - # so this is being maintained for backwards compatibility + result = result_from_cache(filename, result) + + return result + + [email protected]("sourcemaps.get_from_archive") +def get_from_archive(url: str, archive: ReleaseArchive) -> Tuple[bytes, dict]: + candidates = ReleaseFile.normalize(url) + for candidate in candidates: try: - encoding = result[3] + return archive.get_file_by_url(candidate) + except KeyError: + pass + + # None of the filenames matched + raise KeyError(f"Not found in archive: '{url}'") + + [email protected]("sourcemaps.fetch_release_archive") +def fetch_release_archive(release, dist) -> Optional[IO]: + """Fetch release archive and cache if possible. + + If return value is not empty, the caller is responsible for closing the stream. + """ + dist_name = dist and dist.name or None + releasefile_ident = ReleaseFile.get_ident(RELEASE_ARCHIVE_FILENAME, dist_name) + cache_key = get_release_file_cache_key( + release_id=release.id, releasefile_ident=releasefile_ident + ) + + result = cache.get(cache_key) + + if result == -1: + return None + elif result: + return BytesIO(result) + else: + qs = ReleaseFile.objects.filter( + release=release, dist=dist, ident=releasefile_ident + ).select_related("file") + try: + releasefile = qs[0] except IndexError: - encoding = None - result = http.UrlResult( - filename, result[0], zlib.decompress(result[1]), result[2], encoding - ) + # Cache as nonexistent: + cache.set(cache_key, -1, 60) + return None + else: + try: + file_ = fetch_retry_policy(lambda: ReleaseFile.cache.getfile(releasefile)) + except Exception: + logger.error("sourcemaps.read_archive_failed", exc_info=sys.exc_info()) + + return None + + # This will implicitly skip too large payloads. + cache.set(cache_key, file_.read(), 3600) + file_.seek(0) + + return file_ + + +def compress(fp: IO) -> Tuple[bytes, bytes]: + """ Alternative for compress_file when fp does not support chunks """ + content = fp.read() + return zlib.compress(content), content + + +def fetch_release_artifact(url, release, dist): + """ + Get a release artifact either by extracting it or fetching it directly. + + If a release archive was saved, the individual file will be extracted + from the archive. + + """ + cache_key, cache_key_meta = get_cache_keys(url, release, dist) + + result = cache.get(cache_key) + + if result == -1: # Cached as unavailable + return None + + if result: + return result_from_cache(url, result) + + start = time.monotonic() + + release_file = fetch_release_archive(release, dist) + if release_file is not None: + try: + archive = ReleaseArchive(release_file) + except BaseException as exc: + logger.error("Failed to initialize archive for release %s", release.id, exc_info=exc) + # TODO(jjbayer): cache error and return here + else: + with archive: + try: + fp, headers = get_from_archive(url, archive) + except KeyError: + logger.debug( + "Release artifact %r not found in archive (release_id=%s)", url, release.id + ) + cache.set(cache_key, -1, 60) + metrics.timing( + "sourcemaps.release_artifact_from_archive", time.monotonic() - start + ) + return None + except BaseException as exc: + logger.error("Failed to read %s from release %s", url, release.id, exc_info=exc) + # TODO(jjbayer): cache error and return here + else: + result = fetch_and_cache_artifact( + url, + lambda: fp, + cache_key, + cache_key_meta, + headers, + # Cannot use `compress_file` because `ZipExtFile` does not support chunks + compress_fn=compress, + ) + metrics.timing( + "sourcemaps.release_artifact_from_archive", time.monotonic() - start + ) + + return result + + # Fall back to maintain compatibility with old releases and versions of + # sentry-cli which upload files individually + result = fetch_release_file(url, release, dist) return result @@ -350,7 +516,6 @@ def fetch_file(url, project=None, release=None, dist=None, allow_scraping=True): separately, whether or not those attempts are successful. Used for both source files and source maps. """ - # If our url has been truncated, it'd be impossible to fetch # so we check for this early and bail if url[-3:] == "...": @@ -358,7 +523,11 @@ def fetch_file(url, project=None, release=None, dist=None, allow_scraping=True): # if we've got a release to look on, try that first (incl associated cache) if release: - with metrics.timer("sourcemaps.release_file"): + sample_rate = options.get("processing.use-release-archives-sample-rate") + if sample_rate and random.random() < sample_rate: + # Read from archive + result = fetch_release_artifact(url, release, dist) + else: result = fetch_release_file(url, release, dist) else: result = None @@ -489,7 +658,11 @@ def fetch_sourcemap(url, project=None, release=None, dist=None, allow_scraping=T else: # look in the database and, if not found, optionally try to scrape the web result = fetch_file( - url, project=project, release=release, dist=dist, allow_scraping=allow_scraping + url, + project=project, + release=release, + dist=dist, + allow_scraping=allow_scraping, ) body = result.body try: @@ -563,6 +736,7 @@ def __init__(self, *args, **kwargs): self.allow_scraping = organization.get_option( "sentry:scrape_javascript", True ) is not False and self.project.get_option("sentry:scrape_javascript", True) + self.fetch_count = 0 self.sourcemaps_touched = set() @@ -625,7 +799,6 @@ def process_frame(self, processable_frame, processing_task): """ Attempt to demangle the given frame. """ - frame = processable_frame.frame token = None diff --git a/src/sentry/models/file.py b/src/sentry/models/file.py index 46ecb0ce94fd25..f59bc44a5d5715 100644 --- a/src/sentry/models/file.py +++ b/src/sentry/models/file.py @@ -1,3 +1,4 @@ +import io import mmap import os import tempfile @@ -431,9 +432,10 @@ def assemble_from_file_blob_ids(self, file_blob_ids, checksum, commit=True): offset = 0 for blob in file_blobs: FileBlobIndex.objects.create(file=self, blob=blob, offset=offset) - for chunk in blob.getfile().chunks(): - new_checksum.update(chunk) - tf.write(chunk) + with blob.getfile() as blobfile: + for chunk in blobfile.chunks(): + new_checksum.update(chunk) + tf.write(chunk) offset += blob.size self.size = offset @@ -563,7 +565,7 @@ def close(self): self._curidx = None self.closed = True - def seek(self, pos): + def _seek(self, pos): if self.closed: raise ValueError("I/O operation on closed file") @@ -586,6 +588,16 @@ def seek(self, pos): raise ValueError("Cannot seek to pos") self._curfile.seek(pos - self._curidx.offset) + def seek(self, pos, whence=io.SEEK_SET): + if whence == io.SEEK_SET: + return self._seek(pos) + if whence == io.SEEK_CUR: + return self._seek(self.tell() + pos) + if whence == io.SEEK_END: + return self._seek(self.size + pos) + + raise ValueError(f"Invalid value for whence: {whence}") + def tell(self): if self.closed: raise ValueError("I/O operation on closed file") diff --git a/src/sentry/models/releasefile.py b/src/sentry/models/releasefile.py index 102e174fcf8a64..7ae6a586c21a68 100644 --- a/src/sentry/models/releasefile.py +++ b/src/sentry/models/releasefile.py @@ -1,5 +1,8 @@ import errno import os +import zipfile +from tempfile import TemporaryDirectory +from typing import IO, Tuple from urllib.parse import urlsplit, urlunsplit from django.core.files.base import File as FileObj @@ -8,8 +11,9 @@ from sentry import options from sentry.db.models import BoundedPositiveIntegerField, FlexibleForeignKey, Model, sane_repr from sentry.models import clear_cached_files -from sentry.utils import metrics +from sentry.utils import json, metrics from sentry.utils.hashlib import sha1_text +from sentry.utils.zip import safe_extract_zip class ReleaseFile(Model): @@ -118,3 +122,66 @@ def clear_old_entries(self): ReleaseFile.cache = ReleaseFileCache() + + +class ReleaseArchive: + """ Read-only view of uploaded ZIP-archive of release files """ + + def __init__(self, fileobj: IO): + self._fileobj = fileobj + self._zip_file = zipfile.ZipFile(self._fileobj) + self.manifest = self._read_manifest() + files = self.manifest.get("files", {}) + + self._entries_by_url = {entry["url"]: (path, entry) for path, entry in files.items()} + + def __enter__(self): + return self + + def __exit__(self, exc, value, tb): + self._zip_file.close() + self._fileobj.close() + + def read(self, filename: str) -> bytes: + return self._zip_file.read(filename) + + def _read_manifest(self) -> dict: + manifest_bytes = self.read("manifest.json") + return json.loads(manifest_bytes.decode("utf-8")) + + def get_file_by_url(self, url: str) -> Tuple[IO, dict]: + """Return file-like object and headers. + + The caller is responsible for closing the returned stream. + + May raise ``KeyError`` + """ + filename, entry = self._entries_by_url[url] + return self._zip_file.open(filename), entry.get("headers", {}) + + def extract(self) -> TemporaryDirectory: + """Extract contents to a temporary directory. + + The caller is responsible for cleanup of the temporary files. + """ + temp_dir = TemporaryDirectory() + safe_extract_zip(self._fileobj, temp_dir.name, strip_toplevel=False) + + return temp_dir + + +def merge_release_archives(archive1: ReleaseArchive, archive2: ReleaseArchive, target: IO): + """Fields in archive2 take precedence over fields in archive1. """ + merged_manifest = dict(archive1.manifest, **archive2.manifest) + files1 = archive1.manifest.get("files", {}) + files2 = archive2.manifest.get("files", {}) + + merged_manifest["files"] = dict(files1, **files2) + + with zipfile.ZipFile(target, mode="w", compression=zipfile.ZIP_DEFLATED) as zip_file: + for filename in files2.keys(): + zip_file.writestr(filename, archive2.read(filename)) + for filename in files1.keys() - files2.keys(): + zip_file.writestr(filename, archive1.read(filename)) + + zip_file.writestr("manifest.json", json.dumps(merged_manifest)) diff --git a/src/sentry/options/defaults.py b/src/sentry/options/defaults.py index 73d7982d150e64..e1aacfa488b822 100644 --- a/src/sentry/options/defaults.py +++ b/src/sentry/options/defaults.py @@ -336,3 +336,9 @@ # Killswitch for dropping events in symbolicate_event register("store.load-shed-symbolicate-event-projects", type=Any, default=[]) + +# Store release files bundled as zip files +register("processing.save-release-archives", default=False) + +# Try to read release artifacts from zip archives +register("processing.use-release-archives-sample-rate", default=0.0) diff --git a/src/sentry/tasks/assemble.py b/src/sentry/tasks/assemble.py index 083104fa2c5042..44eefce371c595 100644 --- a/src/sentry/tasks/assemble.py +++ b/src/sentry/tasks/assemble.py @@ -1,37 +1,40 @@ import hashlib import logging +from io import BytesIO from os import path from django.db import IntegrityError, transaction +from sentry import options from sentry.api.serializers import serialize from sentry.cache import default_cache +from sentry.models import File, Organization, Release, ReleaseFile +from sentry.models.releasefile import ReleaseArchive, merge_release_archives from sentry.tasks.base import instrumented_task -from sentry.utils import json, metrics +from sentry.utils import metrics from sentry.utils.files import get_max_file_size from sentry.utils.sdk import bind_organization_context, configure_scope logger = logging.getLogger(__name__) -def enum(**named_values): - """Creates an enum type.""" - return type("Enum", (), named_values) +#: Name for the bundle stored as a release file +RELEASE_ARCHIVE_FILENAME = "release-artifacts.zip" +#: How often should we retry merging archives when there's a conflict? +RELEASE_ARCHIVE_MAX_MERGE_ATTEMPTS = 3 -ChunkFileState = enum( - OK="ok", # File in database - NOT_FOUND="not_found", # File not found in database - CREATED="created", # File was created in the request and send to the worker for assembling - ASSEMBLING="assembling", # File still being processed by worker - ERROR="error", # Error happened during assembling -) +class ChunkFileState: + OK = "ok" # File in database + NOT_FOUND = "not_found" # File not found in database + CREATED = "created" # File was created in the request and send to the worker for assembling + ASSEMBLING = "assembling" # File still being processed by worker + ERROR = "error" # Error happened during assembling -AssembleTask = enum( - DIF="project.dsym", # Debug file upload - ARTIFACTS="organization.artifacts", # Release file upload -) +class AssembleTask: + DIF = "project.dsym" # Debug file upload + ARTIFACTS = "organization.artifacts" # Release file upload def _get_cache_key(task, scope, checksum): @@ -85,7 +88,6 @@ def assemble_dif(project_id, name, checksum, chunks, debug_id=None, **kwargs): """ Assembles uploaded chunks into a ``ProjectDebugFile``. """ - from sentry.models import BadDif, Project, debugfile from sentry.reprocessing import bump_reprocessing_revision @@ -163,21 +165,96 @@ class AssembleArtifactsError(Exception): pass +def _simple_update(release_file: ReleaseFile, new_file: File, new_archive: ReleaseArchive): + """ Update function used in _upsert_release_file """ + old_file = release_file.file + release_file.update(file=new_file) + old_file.delete() + + +def _upsert_release_file(file: File, archive: ReleaseArchive, update_fn, **kwargs): + release_file = None + + # Release files must have unique names within their release + # and dist. If a matching file already exists, replace its + # file with the new one; otherwise create it. + try: + release_file = ReleaseFile.objects.get(**kwargs) + except ReleaseFile.DoesNotExist: + try: + with transaction.atomic(): + release_file = ReleaseFile.objects.create(file=file, **kwargs) + except IntegrityError: + # NB: This indicates a race, where another assemble task or + # file upload job has just created a conflicting file. Since + # we're upserting here anyway, yield to the faster actor and + # do not try again. + file.delete() + else: + update_fn(release_file, file, archive) + + [email protected]("tasks.assemble.merge_archives") +def _merge_archives(release_file: ReleaseFile, new_file: File, new_archive: ReleaseArchive): + max_attempts = RELEASE_ARCHIVE_MAX_MERGE_ATTEMPTS + success = False + for attempt in range(max_attempts): + old_file = release_file.file + with ReleaseArchive(old_file.getfile().file) as old_archive: + buffer = BytesIO() + merge_release_archives(old_archive, new_archive, buffer) + + replacement = File.objects.create(name=old_file.name, type=old_file.type) + buffer.seek(0) + replacement.putfile(buffer) + + with transaction.atomic(): + release_file.refresh_from_db() + if release_file.file == old_file: + # Nothing has changed. It is safe to update + release_file.update(file=replacement) + success = True + break + else: + metrics.incr("tasks.assemble.merge_archives_retry", instance=str(attempt)) + else: + logger.error("Failed to merge archive in %s attempts, giving up.", max_attempts) + + if success: + old_file.delete() + + new_file.delete() + + +def _store_single_files(archive: ReleaseArchive, meta: dict): + try: + temp_dir = archive.extract() + except BaseException: + raise AssembleArtifactsError("failed to extract bundle") + + with temp_dir: + artifacts = archive.manifest.get("files", {}) + for rel_path, artifact in artifacts.items(): + artifact_url = artifact.get("url", rel_path) + artifact_basename = artifact_url.rsplit("/", 1)[-1] + + file = File.objects.create( + name=artifact_basename, type="release.file", headers=artifact.get("headers", {}) + ) + + full_path = path.join(temp_dir.name, rel_path) + with open(full_path, "rb") as fp: + file.putfile(fp, logger=logger) + + kwargs = dict(meta, name=artifact_url) + _upsert_release_file(file, None, _simple_update, **kwargs) + + @instrumented_task(name="sentry.tasks.assemble.assemble_artifacts", queue="assemble") def assemble_artifacts(org_id, version, checksum, chunks, **kwargs): """ Creates release files from an uploaded artifact bundle. """ - - import shutil - import tempfile - - from sentry.models import File, Organization, Release, ReleaseFile - from sentry.utils.zip import safe_extract_zip - - scratchpad = None - delete_bundle = False - try: organization = Organization.objects.get_from_cache(pk=org_id) bind_organization_context(organization) @@ -188,7 +265,7 @@ def assemble_artifacts(org_id, version, checksum, chunks, **kwargs): rv = assemble_file( AssembleTask.ARTIFACTS, organization, - "release-artifacts.zip", + RELEASE_ARCHIVE_FILENAME, checksum, chunks, file_type="release.bundle", @@ -201,84 +278,50 @@ def assemble_artifacts(org_id, version, checksum, chunks, **kwargs): return bundle, temp_file = rv - scratchpad = tempfile.mkdtemp() - - # Initially, always delete the bundle file. Later on, we can start to store - # the artifact bundle as a release file. - delete_bundle = True - - try: - safe_extract_zip(temp_file, scratchpad, strip_toplevel=False) - except BaseException: - raise AssembleArtifactsError("failed to extract bundle") try: - manifest_path = path.join(scratchpad, "manifest.json") - with open(manifest_path, "rb") as manifest: - manifest = json.loads(manifest.read()) + archive = ReleaseArchive(temp_file) except BaseException: raise AssembleArtifactsError("failed to open release manifest") - org_slug = manifest.get("org") - if organization.slug != org_slug: - raise AssembleArtifactsError("organization does not match uploaded bundle") - - release_name = manifest.get("release") - if release_name != version: - raise AssembleArtifactsError("release does not match uploaded bundle") - - try: - release = Release.objects.get(organization_id=organization.id, version=release_name) - except Release.DoesNotExist: - raise AssembleArtifactsError("release does not exist") + with archive: + manifest = archive.manifest - dist_name = manifest.get("dist") - dist = None - if dist_name: - dist = release.add_dist(dist_name) + org_slug = manifest.get("org") + if organization.slug != org_slug: + raise AssembleArtifactsError("organization does not match uploaded bundle") - artifacts = manifest.get("files", {}) - for rel_path, artifact in artifacts.items(): - artifact_url = artifact.get("url", rel_path) - artifact_basename = artifact_url.rsplit("/", 1)[-1] + release_name = manifest.get("release") + if release_name != version: + raise AssembleArtifactsError("release does not match uploaded bundle") - file = File.objects.create( - name=artifact_basename, type="release.file", headers=artifact.get("headers", {}) - ) + try: + release = Release.objects.get(organization_id=organization.id, version=release_name) + except Release.DoesNotExist: + raise AssembleArtifactsError("release does not exist") - full_path = path.join(scratchpad, rel_path) - with open(full_path, "rb") as fp: - file.putfile(fp, logger=logger) + dist_name = manifest.get("dist") + dist = None + if dist_name: + dist = release.add_dist(dist_name) - kwargs = { + meta = { # Required for release file creation "organization_id": organization.id, "release": release, - "name": artifact_url, "dist": dist, } + if options.get("processing.save-release-archives"): + kwargs = dict(meta, name=RELEASE_ARCHIVE_FILENAME) + _upsert_release_file(bundle, archive, _merge_archives, **kwargs) - # Release files must have unique names within their release - # and dist. If a matching file already exists, replace its - # file with the new one; otherwise create it. - try: - release_file = ReleaseFile.objects.get(**kwargs) - except ReleaseFile.DoesNotExist: - try: - with transaction.atomic(): - ReleaseFile.objects.create(file=file, **kwargs) - except IntegrityError: - # NB: This indicates a race, where another assemble task or - # file upload job has just created a conflicting file. Since - # we're upserting here anyway, yield to the faster actor and - # do not try again. - file.delete() - else: - old_file = release_file.file - release_file.update(file=file) - old_file.delete() - - # Count files extracted, to compare them to release files endpoint - metrics.incr("tasks.assemble.extracted_files", amount=len(artifacts)) + # NOTE(jjbayer): Single files are still stored to enable + # rolling back from release archives. Once release archives run + # smoothely, this call can be removed / only called when feature + # flag is off. + _store_single_files(archive, meta) + + # Count files extracted, to compare them to release files endpoint + metrics.incr("tasks.assemble.extracted_files", amount=len(manifest.get("files", {}))) except AssembleArtifactsError as e: set_assemble_status( @@ -295,11 +338,6 @@ def assemble_artifacts(org_id, version, checksum, chunks, **kwargs): ) else: set_assemble_status(AssembleTask.ARTIFACTS, org_id, checksum, ChunkFileState.OK) - finally: - if scratchpad: - shutil.rmtree(scratchpad) - if delete_bundle: - bundle.delete() def assemble_file(task, org_or_project, name, checksum, chunks, file_type): diff --git a/src/sentry/testutils/factories.py b/src/sentry/testutils/factories.py index bcd3037ed152ea..c9d5e40b779794 100644 --- a/src/sentry/testutils/factories.py +++ b/src/sentry/testutils/factories.py @@ -213,7 +213,8 @@ def make_word(words=None): def _patch_artifact_manifest(path, org, release, project=None): - manifest = json.loads(open(path, "rb").read()) + with open(path, "rb") as fp: + manifest = json.load(fp) manifest["org"] = org manifest["release"] = release if project: diff --git a/src/sentry/utils/json.py b/src/sentry/utils/json.py index 5b70523e5d33ad..4001ca680ac888 100644 --- a/src/sentry/utils/json.py +++ b/src/sentry/utils/json.py @@ -101,7 +101,7 @@ def dumps(value: JSONData, escape: bool = False, **kwargs) -> str: return _default_encoder.encode(value) -def load(fp, **kwargs) -> str: +def load(fp, **kwargs) -> JSONData: return loads(fp.read()) diff --git a/tests/relay_integration/lang/javascript/test_plugin.py b/tests/relay_integration/lang/javascript/test_plugin.py index 5bd3e5cce5654e..e3b83cb80844d0 100644 --- a/tests/relay_integration/lang/javascript/test_plugin.py +++ b/tests/relay_integration/lang/javascript/test_plugin.py @@ -1,12 +1,16 @@ import os.path +import zipfile from base64 import b64encode +from io import BytesIO import responses from django.utils.encoding import force_bytes from sentry.models import File, Release, ReleaseFile +from sentry.tasks.assemble import RELEASE_ARCHIVE_FILENAME from sentry.testutils import RelayStoreHelper, SnubaTestCase, TransactionTestCase from sentry.testutils.helpers.datetime import before_now, iso_format +from sentry.utils import json from sentry.utils.compat.mock import patch BASE64_SOURCEMAP = "data:application/json;base64," + ( @@ -1111,6 +1115,113 @@ def test_html_response_for_js(self): {"url": "http://example.com/file2.js", "type": "js_invalid_content"}, ] + def _test_expansion_via_release_archive(self, link_sourcemaps: bool): + project = self.project + release = Release.objects.create(organization_id=project.organization_id, version="abc") + release.add_project(project) + + manifest = { + "org": self.organization.slug, + "release": release.version, + "files": { + "files/_/_/file.min.js": { + "url": "http://example.com/file.min.js", + }, + "files/_/_/file1.js": { + "url": "http://example.com/file1.js", + }, + "files/_/_/file2.js": { + "url": "http://example.com/file2.js", + }, + "files/_/_/file.sourcemap.js": { + "url": "http://example.com/file.sourcemap.js", + }, + }, + } + + file_like = BytesIO() + with zipfile.ZipFile(file_like, "w") as zip: + for rel_path, entry in manifest["files"].items(): + name = os.path.basename(rel_path) + content = load_fixture(name) + if name == "file.min.js" and not link_sourcemaps: + # Remove link to source map, add to header instead + content = content.replace(b"//@ sourceMappingURL=file.sourcemap.js", b"") + entry["headers"] = {"SourceMap": "/file.sourcemap.js"} + zip.writestr(rel_path, content) + zip.writestr("manifest.json", json.dumps(manifest)) + file_like.seek(0) + + file = File.objects.create(name=RELEASE_ARCHIVE_FILENAME) + file.putfile(file_like) + + ReleaseFile.objects.create( + name=RELEASE_ARCHIVE_FILENAME, + release=release, + organization_id=project.organization_id, + file=file, + ) + + data = { + "timestamp": self.min_ago, + "message": "hello", + "platform": "javascript", + "release": "abc", + "exception": { + "values": [ + { + "type": "Error", + "stacktrace": { + "frames": [ + { + "abs_path": "http://example.com/file.min.js", + "filename": "file.min.js", + "lineno": 1, + "colno": 39, + }, + { + "abs_path": "http://example.com/file.min.js", + "filename": "file.min.js", + "lineno": 1, + "colno": 79, + }, + ] + }, + } + ] + }, + } + + with self.options({"processing.use-release-archives-sample-rate": 1.1}): + event = self.post_and_retrieve_event(data) + + assert "errors" not in event.data + + exception = event.interfaces["exception"] + frame_list = exception.values[0].stacktrace.frames + + frame = frame_list[0] + assert frame.pre_context == ["function add(a, b) {", '\t"use strict";'] + assert frame.context_line == "\treturn a + b; // fôo" + assert frame.post_context == ["}", ""] + + frame = frame_list[1] + assert frame.pre_context == ["function multiply(a, b) {", '\t"use strict";'] + assert frame.context_line == "\treturn a * b;" + assert frame.post_context == [ + "}", + "function divide(a, b) {", + '\t"use strict";', + "\ttry {", + "\t\treturn multiply(add(a, b), a, b) / c;", + ] + + def test_expansion_via_release_archive(self): + self._test_expansion_via_release_archive(link_sourcemaps=True) + + def test_expansion_via_release_archive_no_sourcemap_link(self): + self._test_expansion_via_release_archive(link_sourcemaps=False) + def test_node_processing(self): project = self.project release = Release.objects.create( diff --git a/tests/sentry/lang/javascript/test_processor.py b/tests/sentry/lang/javascript/test_processor.py index e4de0bead41ec4..ec5471d6138877 100644 --- a/tests/sentry/lang/javascript/test_processor.py +++ b/tests/sentry/lang/javascript/test_processor.py @@ -1,6 +1,7 @@ import errno import re import unittest +import zipfile from copy import deepcopy from io import BytesIO @@ -14,11 +15,13 @@ from sentry.lang.javascript.processor import ( CACHE_CONTROL_MAX, CACHE_CONTROL_MIN, + RELEASE_ARCHIVE_FILENAME, JavaScriptStacktraceProcessor, UnparseableSourcemap, cache, discover_sourcemap, fetch_file, + fetch_release_archive, fetch_release_file, fetch_sourcemap, generate_module, @@ -30,6 +33,7 @@ ) from sentry.models import EventError, File, Release, ReleaseFile from sentry.testutils import TestCase +from sentry.utils import json from sentry.utils.compat.mock import ANY, MagicMock, call, patch from sentry.utils.strings import truncatechars @@ -476,6 +480,7 @@ def test_non_url_without_release(self): @responses.activate @patch("sentry.lang.javascript.processor.fetch_release_file") def test_non_url_with_release(self, mock_fetch_release_file): + mock_fetch_release_file.return_value = http.UrlResult( "/example.js", {"content-type": "application/json"}, b"foo", 200, None ) @@ -490,6 +495,120 @@ def test_non_url_with_release(self, mock_fetch_release_file): assert result.headers == {"content-type": "application/json"} assert result.encoding is None + @responses.activate + def test_non_url_with_release_archive(self): + compressed = BytesIO() + with zipfile.ZipFile(compressed, mode="w") as zip_file: + zip_file.writestr("example.js", b"foo") + zip_file.writestr( + "manifest.json", + json.dumps( + { + "files": { + "example.js": { + "url": "/example.js", + "headers": {"content-type": "application/json"}, + } + } + } + ), + ) + + release = Release.objects.create(version="1", organization_id=self.project.organization_id) + release.add_project(self.project) + + file = File.objects.create( + name=RELEASE_ARCHIVE_FILENAME, + ) + compressed.seek(0) + file.putfile(compressed) + + ReleaseFile.objects.create( + name=RELEASE_ARCHIVE_FILENAME, + release=release, + organization_id=self.project.organization_id, + file=file, + ) + + with self.options({"processing.use-release-archives-sample-rate": 1.0}): + # Attempt to fetch nonexisting + with pytest.raises(http.BadSource): + fetch_file("does-not-exist.js", release=release) + + # Attempt to fetch nonexsting again (to check if cache works) + with pytest.raises(http.BadSource): + result = fetch_file("does-not-exist.js", release=release) + + result = fetch_file("/example.js", release=release) + assert result.url == "/example.js" + assert result.body == b"foo" + assert isinstance(result.body, bytes) + assert result.headers == {"content-type": "application/json"} + assert result.encoding == "utf-8" + + # Make sure cache loading works: + result2 = fetch_file("/example.js", release=release) + assert result2 == result + + @patch("sentry.lang.javascript.processor.cache.set", side_effect=cache.set) + @patch("sentry.lang.javascript.processor.cache.get", side_effect=cache.get) + def test_archive_caching(self, cache_get, cache_set): + release = Release.objects.create(version="1", organization_id=self.project.organization_id) + + def relevant_calls(mock): + return [ + call + for call in mock.mock_calls + if ( + call.args and call.args[0] or call.kwargs and call.kwargs["key"] or "" + ).startswith("releasefile") + ] + + # No archive exists: + result = fetch_release_archive(release, dist=None) + assert result is None + assert len(relevant_calls(cache_get)) == 1 + assert len(relevant_calls(cache_set)) == 1 + cache_get.reset_mock() + cache_set.reset_mock() + + # Still no archive, cache is only read + result = fetch_release_archive(release, dist=None) + assert result is None + assert len(relevant_calls(cache_get)) == 1 + assert len(relevant_calls(cache_set)) == 0 + cache_get.reset_mock() + cache_set.reset_mock() + + file = File.objects.create( + name=RELEASE_ARCHIVE_FILENAME, + ) + file.putfile(BytesIO(b"foo")) + + release = Release.objects.create(version="2", organization_id=self.project.organization_id) + ReleaseFile.objects.create( + name=RELEASE_ARCHIVE_FILENAME, + release=release, + organization_id=self.project.organization_id, + file=file, + ) + + # No we have one, call set again + result = fetch_release_archive(release, dist=None) + assert result not in (None, -1) + assert len(relevant_calls(cache_get)) == 1 + assert len(relevant_calls(cache_set)) == 1 + cache_get.reset_mock() + cache_set.reset_mock() + + # Second time, get it from cache + result = fetch_release_archive(release, dist=None) + assert result not in (None, -1) + assert len(relevant_calls(cache_get)) == 1 + assert len(relevant_calls(cache_set)) == 0 + cache_get.reset_mock() + cache_set.reset_mock() + @responses.activate def test_unicode_body(self): responses.add( diff --git a/tests/sentry/models/test_file.py b/tests/sentry/models/test_file.py index 245d9da7f58a09..478f1463894f8c 100644 --- a/tests/sentry/models/test_file.py +++ b/tests/sentry/models/test_file.py @@ -1,4 +1,5 @@ import os +from io import BytesIO from unittest.mock import patch from django.core.files.base import ContentFile @@ -126,6 +127,34 @@ def test_file_handling(self): with self.assertRaises(ValueError): fp.read() + def test_seek(self): + """ Test behavior of seek with difference values for whence """ + bytes = BytesIO(b"abcdefghijklmnopqrstuvwxyz") + file1 = File.objects.create(name="baz.js", type="default", size=26) + results = file1.putfile(bytes, 5) + assert len(results) == 6 + + with file1.getfile() as fp: + assert fp.read() == b"abcdefghijklmnopqrstuvwxyz" + + fp.seek(0, 2) + bytes.seek(0, 2) + assert fp.tell() == bytes.tell() == 26 + assert fp.read() == bytes.read() == b"" + + fp.seek(-1, 2) + bytes.seek(-1, 2) + assert fp.tell() == bytes.tell() == 25 + assert fp.read() == bytes.read() == b"z" + + fp.seek(-10, 1) + bytes.seek(-10, 1) + assert fp.tell() == bytes.tell() == 16 + assert fp.read() == bytes.read() == b"qrstuvwxyz" + + with self.assertRaises(ValueError): + fp.seek(0, 666) + def test_multi_chunk_prefetch(self): random_data = os.urandom(1 << 25) diff --git a/tests/sentry/models/test_releasefile.py b/tests/sentry/models/test_releasefile.py index d34db922f0b0c4..c0df3fa438d52c 100644 --- a/tests/sentry/models/test_releasefile.py +++ b/tests/sentry/models/test_releasefile.py @@ -1,10 +1,12 @@ import errno import os from io import BytesIO +from zipfile import ZipFile from sentry import options -from sentry.models import ReleaseFile +from sentry.models import ReleaseArchive, ReleaseFile, merge_release_archives from sentry.testutils import TestCase +from sentry.utils import json class ReleaseFileTestCase(TestCase): @@ -82,3 +84,62 @@ def test_getfile_streaming(self): assert e.errno == errno.ENOENT else: assert False, "file should not exist" + + +class ReleaseArchiveTestCase(TestCase): + @staticmethod + def create_archive(fields, files): + manifest = dict( + fields, files={filename: {"url": f"fake://{filename}"} for filename in files} + ) + buffer = BytesIO() + with ZipFile(buffer, mode="w") as zf: + zf.writestr("manifest.json", json.dumps(manifest)) + for filename, content in files.items(): + zf.writestr(filename, content) + + return ReleaseArchive(buffer) + + def test_merge(self): + archive1 = self.create_archive( + fields={ + "org": 1, + "release": 2, + "dist": 3, + }, + files={ + "foo": "foo", + "bar": "bar", + "baz": "baz", + }, + ) + archive2 = self.create_archive( + fields={ + "org": 1, + "release": 666, + "dist": 3, + }, + files={ + "foo": "foo", + "bar": "BAR", + }, + ) + + buffer = BytesIO() + merge_release_archives(archive1, archive2, buffer) + + archive3 = ReleaseArchive(buffer) + + assert archive3.manifest["org"] == 1 + assert archive3.manifest["release"] == 666 + assert archive3.manifest["dist"] == 3 + + assert archive3.manifest["files"].keys() == {"foo", "bar", "baz"} + + # Make sure everything was saved: + peristed_manifest = archive3._read_manifest() + assert peristed_manifest == archive3.manifest + + assert archive3.read("foo") == b"foo" + assert archive3.read("bar") == b"BAR" + assert archive3.read("baz") == b"baz" diff --git a/tests/sentry/tasks/test_assemble.py b/tests/sentry/tasks/test_assemble.py index ed477d3a887351..69bab3b146cf50 100644 --- a/tests/sentry/tasks/test_assemble.py +++ b/tests/sentry/tasks/test_assemble.py @@ -1,14 +1,19 @@ import io import os from hashlib import sha1 +from unittest.mock import patch from django.core.files.base import ContentFile from sentry.models import FileBlob, FileBlobOwner, ReleaseFile from sentry.models.debugfile import ProjectDebugFile +from sentry.models.file import File +from sentry.models.releasefile import ReleaseArchive from sentry.tasks.assemble import ( + RELEASE_ARCHIVE_MAX_MERGE_ATTEMPTS, AssembleTask, ChunkFileState, + _merge_archives, assemble_artifacts, assemble_dif, assemble_file, @@ -186,25 +191,96 @@ def test_artifacts(self): blob1 = FileBlob.from_file(ContentFile(bundle_file)) total_checksum = sha1(bundle_file).hexdigest() - assemble_artifacts( - org_id=self.organization.id, - version=self.release.version, - checksum=total_checksum, - chunks=[blob1.checksum], + for has_release_archives in (True, False): + with self.options({"processing.save-release-archives": has_release_archives}): + + assemble_artifacts( + org_id=self.organization.id, + version=self.release.version, + checksum=total_checksum, + chunks=[blob1.checksum], + ) + + status, details = get_assemble_status( + AssembleTask.ARTIFACTS, self.organization.id, total_checksum + ) + assert status == ChunkFileState.OK + assert details is None + + release_file = ReleaseFile.objects.get( + organization=self.organization, + release=self.release, + name="release-artifacts.zip" if has_release_archives else "~/index.js", + dist=None, + ) + + assert release_file + + if has_release_archives: + assert release_file.file.headers == {} + # Artifact is the same as original bundle + assert release_file.file.size == len(bundle_file) + else: + assert release_file.file.headers == {"Sourcemap": "index.js.map"} + + def test_merge_archives(self): + file1 = File.objects.create() + file1.putfile(ContentFile(self.create_artifact_bundle())) + file2 = File.objects.create() + file2.putfile(ContentFile(self.create_artifact_bundle())) + + release_file = ReleaseFile.objects.create( + organization=self.organization, + release=self.release, + file=file1, ) - status, details = get_assemble_status( - AssembleTask.ARTIFACTS, self.organization.id, total_checksum - ) - assert status == ChunkFileState.OK - assert details is None - - release_file = ReleaseFile.objects.get( - organization=self.organization, release=self.release, name="~/index.js", dist=None + with ReleaseArchive(file2.getfile().file) as archive2: + _merge_archives(release_file, file2, archive2) + # Both files have disappeared, a new one has taken their place: + assert not File.objects.filter(pk=file1.pk).exists() + assert not File.objects.filter(pk=file2.pk).exists() + assert release_file.file.pk > 2 + + @patch("sentry.tasks.assemble.logger.error") + @patch("sentry.tasks.assemble.ReleaseFile.refresh_from_db") + @patch("sentry.tasks.assemble.metrics.incr") + def test_merge_archives_fail(self, mock_incr, mock_refresh, mock_log_error): + max_attempts = RELEASE_ARCHIVE_MAX_MERGE_ATTEMPTS + + file1 = File.objects.create() + file1.putfile(ContentFile(self.create_artifact_bundle())) + file2 = File.objects.create() + file2.putfile(ContentFile(self.create_artifact_bundle())) + + release_file = ReleaseFile.objects.create( + organization=self.organization, + release=self.release, + file=file1, ) - assert release_file - assert release_file.file.headers == {"Sourcemap": "index.js.map"} + with ReleaseArchive(file2.getfile().file) as archive2: + + def change_file_id(): + # Create another file + release_file.file = File.objects.create() + release_file.file.putfile(ContentFile(self.create_artifact_bundle())) + + mock_refresh.side_effect = change_file_id + + _merge_archives(release_file, file2, archive2) + assert mock_refresh.called + # Failed to update + assert File.objects.filter(pk=file1.pk).exists() + assert ReleaseFile.objects.get(pk=release_file.pk).file == file1 + assert not File.objects.filter(pk=file2.pk).exists() + for attempt in range(max_attempts): + assert mock_incr.called_with( + "tasks.assemble.merge_archives_retry", instance=str(attempt) + ) + assert mock_log_error.called_with( + f"Failed to merge archive in {max_attempts} attempts, giving up." + ) def test_artifacts_invalid_org(self): bundle_file = self.create_artifact_bundle(org="invalid")
86cf81d68e5ac709fdf1aea26fb80d6a7a5a0e12
2019-01-25 03:43:05
Lyn Nagara
feat(settings): Update project settings links (#11675)
false
Update project settings links (#11675)
feat
diff --git a/src/sentry/static/sentry/app/views/settings/components/settingsProjectItem.jsx b/src/sentry/static/sentry/app/views/settings/components/settingsProjectItem.jsx index d5ebd825a7886e..760f4052aef905 100644 --- a/src/sentry/static/sentry/app/views/settings/components/settingsProjectItem.jsx +++ b/src/sentry/static/sentry/app/views/settings/components/settingsProjectItem.jsx @@ -54,6 +54,8 @@ const ProjectItem = createReactClass({ let {project, organization} = this.props; let {isBookmarked} = this.state; + let hasNewRoutes = new Set(organization.features).has('sentry10'); + return ( <div key={project.id} className={isBookmarked ? 'isBookmarked' : null}> <Tooltip title={isBookmarked ? 'Remove from bookmarks' : 'Add to bookmarks'}> @@ -65,7 +67,13 @@ const ProjectItem = createReactClass({ )} </InlineButton> </Tooltip> - <Link to={`/${organization.slug}/${project.slug}/`}> + <Link + to={ + hasNewRoutes + ? `/settings/${organization.slug}/${project.slug}/` + : `/${organization.slug}/${project.slug}/` + } + > <ProjectLabel project={project} /> </Link> </div> diff --git a/src/sentry/static/sentry/app/views/settings/organizationProjects/index.jsx b/src/sentry/static/sentry/app/views/settings/organizationProjects/index.jsx index 3f398ffc8689ed..f8e1ed634000ef 100644 --- a/src/sentry/static/sentry/app/views/settings/organizationProjects/index.jsx +++ b/src/sentry/static/sentry/app/views/settings/organizationProjects/index.jsx @@ -75,6 +75,8 @@ export default class OrganizationProjects extends AsyncView { .getAccess() .has('project:admin'); + let hasNewRoutes = new Set(organization.features).has('sentry10'); + let action = ( <Button priority="primary" @@ -121,15 +123,17 @@ export default class OrganizationProjects extends AsyncView { stats={projectStats[project.id]} /> </Box> - <Box p={2} align="right"> - <Button - icon="icon-settings" - size="small" - to={`/settings/${organization.slug}/${project.slug}/`} - > - {t('Settings')} - </Button> - </Box> + {!hasNewRoutes && ( + <Box p={2} align="right"> + <Button + icon="icon-settings" + size="small" + to={`/settings/${organization.slug}/${project.slug}/`} + > + {t('Settings')} + </Button> + </Box> + )} </PanelItem> ))} {projectList.length === 0 && (
22a0c41f432eadf9abc2cd03c29ad178b6247635
2022-09-27 20:25:20
Armen Zambrano G
fix(gitlab): Fix AppPlatformEvents hitting Gitlab webhook endpoint (#39333)
false
Fix AppPlatformEvents hitting Gitlab webhook endpoint (#39333)
fix
diff --git a/src/sentry/integrations/gitlab/webhooks.py b/src/sentry/integrations/gitlab/webhooks.py index 1d1f8bf97e41d0..699bba2e6770f2 100644 --- a/src/sentry/integrations/gitlab/webhooks.py +++ b/src/sentry/integrations/gitlab/webhooks.py @@ -196,7 +196,9 @@ def post(self, request: Request) -> Response: extra = { # This tells us the Gitlab version being used (e.g. current gitlab.com version -> GitLab/15.4.0-pre) "user-agent": request.META.get("HTTP_USER_AGENT"), - "event-type": request.META["HTTP_X_GITLAB_EVENT"], + # Gitlab does not seem to be the only host sending events + # AppPlatformEvents also hit this API + "event-type": request.META.get("HTTP_X_GITLAB_EVENT"), } token = "<unknown>" try:
07da0859ce1fcc2b217cafb146bbb06e78df2e30
2019-10-10 23:20:50
Lyn Nagara
test: Do not test legacy models (#15013)
false
Do not test legacy models (#15013)
test
diff --git a/tests/sentry/db/test_parse_query.py b/tests/sentry/db/test_parse_query.py index e310b4bbcdb16a..8bc064f0295062 100644 --- a/tests/sentry/db/test_parse_query.py +++ b/tests/sentry/db/test_parse_query.py @@ -74,108 +74,8 @@ def test_parse_postgres_queries(self): u"time": u"0.000", }, {u"sql": u'RELEASE SAVEPOINT "s47890194282880_x52"', u"time": u"0.000"}, - { - u"sql": u'SELECT "sentry_filterkey"."id", "sentry_filterkey"."project_id", "sentry_filterkey"."key", "sentry_filterkey"."values_seen", "sentry_filterkey"."label", "sentry_filterkey"."status" FROM "sentry_filterkey" WHERE ("sentry_filterkey"."project_id" = 864 AND "sentry_filterkey"."key" = \'level\' )', - u"time": u"0.001", - }, - { - u"sql": u'SELECT "sentry_filtervalue"."id", "sentry_filtervalue"."project_id", "sentry_filtervalue"."key", "sentry_filtervalue"."value", "sentry_filtervalue"."data", "sentry_filtervalue"."times_seen", "sentry_filtervalue"."last_seen", "sentry_filtervalue"."first_seen" FROM "sentry_filtervalue" WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'error\' AND "sentry_filtervalue"."key" = \'level\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filterkey"."id", "sentry_filterkey"."project_id", "sentry_filterkey"."key", "sentry_filterkey"."values_seen", "sentry_filterkey"."label", "sentry_filterkey"."status" FROM "sentry_filterkey" WHERE ("sentry_filterkey"."project_id" = 864 AND "sentry_filterkey"."key" = \'url\' )', - u"time": u"0.001", - }, - { - u"sql": u'SELECT "sentry_filtervalue"."id", "sentry_filtervalue"."project_id", "sentry_filtervalue"."key", "sentry_filtervalue"."value", "sentry_filtervalue"."data", "sentry_filtervalue"."times_seen", "sentry_filtervalue"."last_seen", "sentry_filtervalue"."first_seen" FROM "sentry_filtervalue" WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'http://example.com\' AND "sentry_filtervalue"."key" = \'url\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filterkey"."id", "sentry_filterkey"."project_id", "sentry_filterkey"."key", "sentry_filterkey"."values_seen", "sentry_filterkey"."label", "sentry_filterkey"."status" FROM "sentry_filterkey" WHERE ("sentry_filterkey"."project_id" = 864 AND "sentry_filterkey"."key" = \'sentry:user\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filtervalue"."id", "sentry_filtervalue"."project_id", "sentry_filtervalue"."key", "sentry_filtervalue"."value", "sentry_filtervalue"."data", "sentry_filtervalue"."times_seen", "sentry_filtervalue"."last_seen", "sentry_filtervalue"."first_seen" FROM "sentry_filtervalue" WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'ip:127.0.0.1\' AND "sentry_filtervalue"."key" = \'sentry:user\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filterkey"."id", "sentry_filterkey"."project_id", "sentry_filterkey"."key", "sentry_filterkey"."values_seen", "sentry_filterkey"."label", "sentry_filterkey"."status" FROM "sentry_filterkey" WHERE ("sentry_filterkey"."project_id" = 864 AND "sentry_filterkey"."key" = \'os.name\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filtervalue"."id", "sentry_filtervalue"."project_id", "sentry_filtervalue"."key", "sentry_filtervalue"."value", "sentry_filtervalue"."data", "sentry_filtervalue"."times_seen", "sentry_filtervalue"."last_seen", "sentry_filtervalue"."first_seen" FROM "sentry_filtervalue" WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'Windows 8\' AND "sentry_filtervalue"."key" = \'os.name\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filterkey"."id", "sentry_filterkey"."project_id", "sentry_filterkey"."key", "sentry_filterkey"."values_seen", "sentry_filterkey"."label", "sentry_filterkey"."status" FROM "sentry_filterkey" WHERE ("sentry_filterkey"."project_id" = 864 AND "sentry_filterkey"."key" = \'browser.name\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filtervalue"."id", "sentry_filtervalue"."project_id", "sentry_filtervalue"."key", "sentry_filtervalue"."value", "sentry_filtervalue"."data", "sentry_filtervalue"."times_seen", "sentry_filtervalue"."last_seen", "sentry_filtervalue"."first_seen" FROM "sentry_filtervalue" WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'Chrome\' AND "sentry_filtervalue"."key" = \'browser.name\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filterkey"."id", "sentry_filterkey"."project_id", "sentry_filterkey"."key", "sentry_filterkey"."values_seen", "sentry_filterkey"."label", "sentry_filterkey"."status" FROM "sentry_filterkey" WHERE ("sentry_filterkey"."project_id" = 864 AND "sentry_filterkey"."key" = \'browser\' )', - u"time": u"0.000", - }, - { - u"sql": u'SELECT "sentry_filtervalue"."id", "sentry_filtervalue"."project_id", "sentry_filtervalue"."key", "sentry_filtervalue"."value", "sentry_filtervalue"."data", "sentry_filtervalue"."times_seen", "sentry_filtervalue"."last_seen", "sentry_filtervalue"."first_seen" FROM "sentry_filtervalue" WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'Chrome 28.0.1500\' AND "sentry_filtervalue"."key" = \'browser\' )', - u"time": u"0.000", - }, {u"sql": u'SAVEPOINT "s47890194282880_x54"', u"time": u"0.000"}, - { - u"sql": u'INSERT INTO "sentry_eventtag" ("project_id", "group_id", "event_id", "key_id", "value_id", "date_added") VALUES (864, 662, 454, 108, 108, \'2018-05-22 09:12:12+00:00\'), (864, 662, 454, 109, 109, \'2018-05-22 09:12:12+00:00\'), (864, 662, 454, 110, 110, \'2018-05-22 09:12:12+00:00\'), (864, 662, 454, 111, 111, \'2018-05-22 09:12:12+00:00\'), (864, 662, 454, 112, 112, \'2018-05-22 09:12:12+00:00\'), (864, 662, 454, 113, 113, \'2018-05-22 09:12:12+00:00\')', - u"time": u"0.000", - }, {u"sql": u'RELEASE SAVEPOINT "s47890194282880_x54"', u"time": u"0.000"}, - { - u"sql": u'UPDATE "sentry_filtervalue" SET "times_seen" = "sentry_filtervalue"."times_seen" + 1, "data" = NULL, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'error\' AND "sentry_filtervalue"."key" = \'level\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_messagefiltervalue" SET "times_seen" = "sentry_messagefiltervalue"."times_seen" + 1, "project_id" = 864, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_messagefiltervalue"."group_id" = 662 AND "sentry_messagefiltervalue"."value" = \'error\' AND "sentry_messagefiltervalue"."key" = \'level\' )', - u"time": u"0.000", - }, - { - u"sql": u'UPDATE "sentry_filtervalue" SET "times_seen" = "sentry_filtervalue"."times_seen" + 1, "data" = NULL, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'http://example.com\' AND "sentry_filtervalue"."key" = \'url\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_messagefiltervalue" SET "times_seen" = "sentry_messagefiltervalue"."times_seen" + 1, "project_id" = 864, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_messagefiltervalue"."group_id" = 662 AND "sentry_messagefiltervalue"."value" = \'http://example.com\' AND "sentry_messagefiltervalue"."key" = \'url\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_filtervalue" SET "times_seen" = "sentry_filtervalue"."times_seen" + 1, "data" = NULL, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'ip:127.0.0.1\' AND "sentry_filtervalue"."key" = \'sentry:user\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_messagefiltervalue" SET "times_seen" = "sentry_messagefiltervalue"."times_seen" + 1, "project_id" = 864, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_messagefiltervalue"."group_id" = 662 AND "sentry_messagefiltervalue"."value" = \'ip:127.0.0.1\' AND "sentry_messagefiltervalue"."key" = \'sentry:user\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_filtervalue" SET "times_seen" = "sentry_filtervalue"."times_seen" + 1, "data" = NULL, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'Windows 8\' AND "sentry_filtervalue"."key" = \'os.name\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_messagefiltervalue" SET "times_seen" = "sentry_messagefiltervalue"."times_seen" + 1, "project_id" = 864, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_messagefiltervalue"."group_id" = 662 AND "sentry_messagefiltervalue"."value" = \'Windows 8\' AND "sentry_messagefiltervalue"."key" = \'os.name\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_filtervalue" SET "times_seen" = "sentry_filtervalue"."times_seen" + 1, "data" = NULL, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'Chrome\' AND "sentry_filtervalue"."key" = \'browser.name\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_messagefiltervalue" SET "times_seen" = "sentry_messagefiltervalue"."times_seen" + 1, "project_id" = 864, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_messagefiltervalue"."group_id" = 662 AND "sentry_messagefiltervalue"."value" = \'Chrome\' AND "sentry_messagefiltervalue"."key" = \'browser.name\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_filtervalue" SET "times_seen" = "sentry_filtervalue"."times_seen" + 1, "data" = NULL, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_filtervalue"."project_id" = 864 AND "sentry_filtervalue"."value" = \'Chrome 28.0.1500\' AND "sentry_filtervalue"."key" = \'browser\' )', - u"time": u"0.001", - }, - { - u"sql": u'UPDATE "sentry_messagefiltervalue" SET "times_seen" = "sentry_messagefiltervalue"."times_seen" + 1, "project_id" = 864, "last_seen" = \'2018-05-22 09:12:12+00:00\' WHERE ("sentry_messagefiltervalue"."group_id" = 662 AND "sentry_messagefiltervalue"."value" = \'Chrome 28.0.1500\' AND "sentry_messagefiltervalue"."key" = \'browser\' )', - u"time": u"0.001", - }, { u"sql": u'SELECT "sentry_groupedmessage"."id", "sentry_groupedmessage"."project_id", "sentry_groupedmessage"."logger", "sentry_groupedmessage"."level", "sentry_groupedmessage"."message", "sentry_groupedmessage"."view", "sentry_groupedmessage"."num_comments", "sentry_groupedmessage"."platform", "sentry_groupedmessage"."status", "sentry_groupedmessage"."times_seen", "sentry_groupedmessage"."last_seen", "sentry_groupedmessage"."first_seen", "sentry_groupedmessage"."first_release_id", "sentry_groupedmessage"."resolved_at", "sentry_groupedmessage"."active_at", "sentry_groupedmessage"."time_spent_total", "sentry_groupedmessage"."time_spent_count", "sentry_groupedmessage"."score", "sentry_groupedmessage"."is_public", "sentry_groupedmessage"."data", "sentry_groupedmessage"."short_id" FROM "sentry_groupedmessage" WHERE "sentry_groupedmessage"."id" = 662 ', u"time": u"0.001", @@ -196,11 +96,8 @@ def test_parse_postgres_queries(self): assert result == { "nodestore_node": 2, "sentry_environmentproject": 1, - "sentry_eventtag": 1, "sentry_eventuser": 1, - "sentry_filtervalue": 6, "sentry_groupedmessage": 1, "sentry_message": 1, - "sentry_messagefiltervalue": 6, "sentry_userreport": 1, }
4ed280d477bbb52361aad743c9f2a7dca87cad32
2021-12-01 22:21:48
Priscila Oliveira
ref(dynamic-sampling): Add feature guard for error rules on server side [INGEST-708] (#30316)
false
Add feature guard for error rules on server side [INGEST-708] (#30316)
ref
diff --git a/src/sentry/api/endpoints/project_details.py b/src/sentry/api/endpoints/project_details.py index d8d8bc92b8a728..f464f526a86743 100644 --- a/src/sentry/api/endpoints/project_details.py +++ b/src/sentry/api/endpoints/project_details.py @@ -431,6 +431,12 @@ def put(self, request, project): "organizations:filters-and-sampling", project.organization, actor=request.user ) + allow_dynamic_sampling_error_rules = features.has( + "organizations:filters-and-sampling-error-rules", + project.organization, + actor=request.user, + ) + if not allow_dynamic_sampling and result.get("dynamicSampling"): # trying to set dynamic sampling with feature disabled return Response( @@ -612,6 +618,19 @@ def put(self, request, project): if "dynamicSampling" in result: raw_dynamic_sampling = result["dynamicSampling"] + if ( + not allow_dynamic_sampling_error_rules + and self._dynamic_sampling_contains_error_rule(raw_dynamic_sampling) + ): + return Response( + { + "detail": [ + "Dynamic Sampling only accepts rules of type transaction or trace" + ] + }, + status=400, + ) + fixed_rules = self._fix_rule_ids(project, raw_dynamic_sampling) project.update_option("sentry:dynamic_sampling", fixed_rules) @@ -814,3 +833,10 @@ def _fix_rule_ids(self, project, raw_dynamic_sampling): raw_dynamic_sampling["next_id"] = next_id return raw_dynamic_sampling + + def _dynamic_sampling_contains_error_rule(self, raw_dynamic_sampling): + if raw_dynamic_sampling is not None: + rules = raw_dynamic_sampling.get("rules", []) + for rule in rules: + if rule["type"] == "error": + return True diff --git a/static/app/views/settings/project/filtersAndSampling/modal/handleXhrErrorResponse.tsx b/static/app/views/settings/project/filtersAndSampling/modal/handleXhrErrorResponse.tsx deleted file mode 100644 index 8e76559f012106..00000000000000 --- a/static/app/views/settings/project/filtersAndSampling/modal/handleXhrErrorResponse.tsx +++ /dev/null @@ -1,39 +0,0 @@ -import {t} from 'sentry/locale'; -import {DynamicSamplingRule} from 'sentry/types/dynamicSampling'; - -type Error = { - type: 'sampleRate' | 'unknown'; - message: string; -}; - -type XhrError = { - responseJSON?: { - dynamicSampling?: { - rules: Array<Partial<DynamicSamplingRule>>; - }; - }; -}; - -function handleXhrErrorResponse(error: XhrError, currentRuleIndex: number): Error { - const responseErrors = - error.responseJSON?.dynamicSampling?.rules[currentRuleIndex] ?? {}; - - const [type, value] = Object.entries(responseErrors)[0]; - - if (type === 'sampleRate') { - const message = Array.isArray(value) ? value[0] : value; - if (message === 'Ensure this value is less than or equal to 1.') { - return { - type: 'sampleRate', - message: t('Ensure this value is a floating number between 0 and 100'), - }; - } - } - - return { - type: 'unknown', - message: t('An internal error occurred while saving dynamic sampling rule'), - }; -} - -export default handleXhrErrorResponse; diff --git a/static/app/views/settings/project/filtersAndSampling/modal/ruleModal.tsx b/static/app/views/settings/project/filtersAndSampling/modal/ruleModal.tsx index dd02fc81c64122..220cc490088581 100644 --- a/static/app/views/settings/project/filtersAndSampling/modal/ruleModal.tsx +++ b/static/app/views/settings/project/filtersAndSampling/modal/ruleModal.tsx @@ -25,8 +25,7 @@ import EmptyMessage from 'sentry/views/settings/components/emptyMessage'; import NumberField from 'sentry/views/settings/components/forms/numberField'; import Conditions from './conditions'; -import handleXhrErrorResponse from './handleXhrErrorResponse'; -import {isLegacyBrowser} from './utils'; +import {getErrorMessage, isLegacyBrowser} from './utils'; type ConditionsProps = React.ComponentProps<typeof Conditions>['conditions']; @@ -138,11 +137,16 @@ function RuleModal({ ); closeModal(); } catch (error) { - convertErrorXhrResponse(handleXhrErrorResponse(error, currentRuleIndex)); + convertRequestErrorResponse(getErrorMessage(error, currentRuleIndex)); } } - function convertErrorXhrResponse(error: ReturnType<typeof handleXhrErrorResponse>) { + function convertRequestErrorResponse(error: ReturnType<typeof getErrorMessage>) { + if (typeof error === 'string') { + addErrorMessage(error); + return; + } + switch (error.type) { case 'sampleRate': setData({...data, errors: {...errors, sampleRate: error.message}}); diff --git a/static/app/views/settings/project/filtersAndSampling/modal/utils.tsx b/static/app/views/settings/project/filtersAndSampling/modal/utils.tsx index cbc86c0c2b5d1c..21a99cafab94e9 100644 --- a/static/app/views/settings/project/filtersAndSampling/modal/utils.tsx +++ b/static/app/views/settings/project/filtersAndSampling/modal/utils.tsx @@ -6,6 +6,7 @@ import { DynamicSamplingConditionLogicalInner, DynamicSamplingInnerName, DynamicSamplingInnerOperator, + DynamicSamplingRule, LegacyBrowser, } from 'sentry/types/dynamicSampling'; import theme from 'sentry/utils/theme'; @@ -156,3 +157,51 @@ export function getNewCondition( }, }; } + +const unexpectedErrorMessage = t( + 'An internal error occurred while saving dynamic sampling rule' +); + +type ResponseJSONDetailed = { + detail: string[]; +}; + +type ResponseJSON = { + dynamicSampling?: { + rules: Array<Partial<DynamicSamplingRule>>; + }; +}; + +export function getErrorMessage( + error: { + responseJSON?: ResponseJSON | ResponseJSONDetailed; + }, + currentRuleIndex: number +) { + const detailedErrorResponse = (error.responseJSON as undefined | ResponseJSONDetailed) + ?.detail; + + if (detailedErrorResponse) { + // This is a temp solution until we enable error rules again, therefore it does not need translation + return detailedErrorResponse[0]; + } + + const errorResponse = error.responseJSON as undefined | ResponseJSON; + + if (!errorResponse) { + return unexpectedErrorMessage; + } + + const responseErrors = errorResponse.dynamicSampling?.rules[currentRuleIndex] ?? {}; + + const [type, _value] = Object.entries(responseErrors)[0]; + + if (type === 'sampleRate') { + return { + type: 'sampleRate', + message: t('Ensure this value is a floating number between 0 and 100'), + }; + } + + return unexpectedErrorMessage; +}
2750139a48bfb295587cdaeecfffc94122aaa984
2018-03-13 01:03:28
Billy Vong
feat(ui): Add default forms to account notification fine tuning (#7567)
false
Add default forms to account notification fine tuning (#7567)
feat
diff --git a/src/sentry/static/sentry/app/data/forms/accountNotificationSettings.jsx b/src/sentry/static/sentry/app/data/forms/accountNotificationSettings.jsx index 31cdcb7300dcdf..9b5786dba7c0f9 100644 --- a/src/sentry/static/sentry/app/data/forms/accountNotificationSettings.jsx +++ b/src/sentry/static/sentry/app/data/forms/accountNotificationSettings.jsx @@ -3,33 +3,60 @@ import {t, tct} from '../../locale'; // Export route to make these forms searchable by label/help export const route = '/settings/account/notifications/'; +export const fields = { + subscribeByDefault: { + name: 'subscribeByDefault', + type: 'boolean', + label: t('Send Me Project Alerts'), + // TODO(billy): Make this a real link + help: tct('Alerts are defined in [locationPath]', { + locationPath: '[Project] » Project Settings » Alerts » Rules.', + }), + }, + workflowNotifications: { + name: 'workflowNotifications', + type: 'radio', + label: t('Send Me Workflow Notifications'), + choices: [[0, 'Always'], [1, 'Only On Issues I Subscribe To'], [2, 'Never']], + help: t('E.g. changes in issue assignment, resolution status, and comments.'), + }, + weeklyReports: { + name: 'weeklyReports', + type: 'boolean', + label: t('Send Me Weekly Reports'), + help: t("Reports contain a summary of what's happened within your organization."), + disabled: true, + }, + deployNotifications: { + name: 'deployNotifications', + type: 'radio', + label: t('Send Me Deploy Notifications'), + choices: [[2, 'Always'], [3, 'Only On Deploys With My Commits'], [4, 'Never']], + help: t('Deploy emails include release, environment and commit overviews.'), + }, + personalActivityNotifications: { + name: 'personalActivityNotifications', + type: 'boolean', + label: t('Notify Me About My Own Activity'), + help: t('Enable this to recieve notifications about your own actions on Sentry.'), + }, + selfAssignOnResolve: { + name: 'selfAssignOnResolve', + type: 'boolean', + label: t("Claim Unassigned Issues I've Resolved"), + help: t("You'll recieve notifications about any changes that happen afterwards."), + }, +}; + const formGroups = [ { title: t('Alerts'), - fields: [ - { - name: 'subscribeByDefault', - type: 'boolean', - label: t('Send Me Project Alerts'), - // TODO(billy): Make this a real link - help: tct('Alerts are defined in [locationPath]', { - locationPath: '[Project] » Project Settings » Alerts » Rules.', - }), - }, - ], + fields: [fields.subscribeByDefault], }, { title: t('Workflow Notifications'), - fields: [ - { - name: 'workflowNotifications', - type: 'radio', - label: t('Send Me Workflow Notifications'), - choices: [[0, 'Always'], [1, 'Only On Issues I Subscribe To'], [2, 'Never']], - help: t('E.g. changes in issue assignment, resolution status, and comments.'), - }, - ], + fields: [fields.workflowNotifications], }, { @@ -39,46 +66,17 @@ const formGroups = [ { title: t('Weekly Reports'), - fields: [ - { - name: 'weeklyReports', - type: 'boolean', - label: t('Send Me Weekly Reports'), - help: t("Reports contain a summary of what's happened within your organization."), - disabled: true, - }, - ], + fields: [fields.weeklyReports], }, { title: t('Deploy Notifications'), - fields: [ - { - name: 'deployNotifications', - type: 'radio', - label: t('Send Me Deploy Notifications'), - choices: [[2, 'Always'], [3, 'Only On Deploys With My Commits'], [4, 'Never']], - help: t('Deploy emails include release, environment and commit overviews.'), - }, - ], + fields: [fields.deployNotifications], }, { title: t('My Activity'), - fields: [ - { - name: 'personalActivityNotifications', - type: 'boolean', - label: t('Notify Me About My Own Activity'), - help: t('Enable this to recieve notifications about your own actions on Sentry.'), - }, - { - name: 'selfAssignOnResolve', - type: 'boolean', - label: t("Claim Unassigned Issues I've Resolved"), - help: t("You'll recieve notifications about any changes that happen afterwards."), - }, - ], + fields: [fields.personalActivityNotifications, fields.selfAssignOnResolve], }, ]; diff --git a/src/sentry/static/sentry/app/views/settings/account/accountNotificationFineTuning.jsx b/src/sentry/static/sentry/app/views/settings/account/accountNotificationFineTuning.jsx index a4007a5a6fbdda..3fb57e4e7f7379 100644 --- a/src/sentry/static/sentry/app/views/settings/account/accountNotificationFineTuning.jsx +++ b/src/sentry/static/sentry/app/views/settings/account/accountNotificationFineTuning.jsx @@ -5,6 +5,7 @@ import styled from 'react-emotion'; import {t} from '../../../locale'; import AsyncView from '../../asyncView'; import Form from '../components/forms/form'; +import JsonForm from '../components/forms/jsonForm'; import Panel from '../components/panel'; import PanelBody from '../components/panelBody'; import PanelHeader from '../components/panelHeader'; @@ -12,6 +13,7 @@ import ProjectsStore from '../../../stores/projectsStore'; import Select2Field from '../components/forms/select2Field'; import SettingsPageHeader from '../components/settingsPageHeader'; import TextBlock from '../components/text/textBlock'; +import {fields} from '../../../data/forms/accountNotificationSettings'; import withOrganizations from '../../../utils/withOrganizations'; const ACCOUNT_NOTIFICATION_FIELDS = { @@ -20,6 +22,7 @@ const ACCOUNT_NOTIFICATION_FIELDS = { description: t('Control alerts that you receive per project.'), type: 'select', choices: [[-1, t('Default')], [1, t('On')], [0, t('Off')]], + defaultFieldName: 'subscribeByDefault', }, workflow: { title: 'Workflow Notifications', @@ -33,6 +36,7 @@ const ACCOUNT_NOTIFICATION_FIELDS = { [1, t('Only on issues I subscribe to')], [2, t('Never')], ], + defaultFieldName: 'workflowNotifications', }, deploy: { title: t('Deploy Notifications'), @@ -46,6 +50,7 @@ const ACCOUNT_NOTIFICATION_FIELDS = { [3, t('Only on deploys with my commits')], [4, t('Never')], ], + defaultFieldName: 'deployNotifications', }, reports: { title: t('Weekly Reports'), @@ -54,6 +59,7 @@ const ACCOUNT_NOTIFICATION_FIELDS = { ), type: 'select', choices: [[1, t('On')], [0, t('Off')]], + defaultFieldName: 'weeklyReports', }, email: { @@ -156,12 +162,12 @@ class AccountNotificationsByOrganization extends React.Component { } render() { - const fields = this.getFieldData(); + const orgFields = this.getFieldData(); return ( <React.Fragment> <PanelHeader>{t('Organizations')}</PanelHeader> - {fields.map(field => { + {orgFields.map(field => { return ( <PanelBodyLineItem key={field.name}> <Select2Field @@ -185,7 +191,10 @@ const AccountNotificationsByOrganizationContainer = withOrganizations( export default class AccountNotificationFineTuning extends AsyncView { getEndpoints() { const {fineTuneType} = this.props.params; - const endpoints = [['notifications', `/users/me/notifications/${fineTuneType}/`]]; + const endpoints = [ + ['notifications', '/users/me/notifications/'], + ['fineTuneData', `/users/me/notifications/${fineTuneType}/`], + ]; if (isGroupedByProject(fineTuneType)) { endpoints.push(['projects', '/projects/']); @@ -232,11 +241,25 @@ export default class AccountNotificationFineTuning extends AsyncView { <SettingsPageHeader title={title} /> {description && <TextBlock>{description}</TextBlock>} + {field && + field.defaultFieldName && ( + <Form + saveOnBlur + apiMethod="PUT" + apiEndpoint={'/users/me/notifications/'} + initialData={this.state.notifications} + > + <JsonForm + title={`Default ${title}`} + fields={[fields[field.defaultFieldName]]} + /> + </Form> + )} <Form saveOnBlur apiMethod="PUT" apiEndpoint={`/users/me/notifications/${this.props.params.fineTuneType}/`} - initialData={this.state.notifications} + initialData={this.state.fineTuneData} > <Panel> {isProject && (
880bb1fc58b681f69bc7332624315da25c477a42
2024-11-06 02:21:56
Nar Saynorath
fix(alerts): Remove extra paren (#80277)
false
Remove extra paren (#80277)
fix
diff --git a/static/app/views/alerts/rules/metric/triggers/chart/index.tsx b/static/app/views/alerts/rules/metric/triggers/chart/index.tsx index baf73a7f7b96d8..fd3f7de98c0175 100644 --- a/static/app/views/alerts/rules/metric/triggers/chart/index.tsx +++ b/static/app/views/alerts/rules/metric/triggers/chart/index.tsx @@ -540,7 +540,6 @@ class TriggersChart extends PureComponent<Props, State> { }); }} </OnDemandMetricRequest> - ); </Fragment> ); }
5638b9eb2c606f9d30285e69793bd8ab8464a2b2
2023-11-17 03:53:57
Stephen Cefali
ref(notifications): remove most instances of should_use_notifications_v2 (#59810)
false
remove most instances of should_use_notifications_v2 (#59810)
ref
diff --git a/src/sentry/api/serializers/models/group.py b/src/sentry/api/serializers/models/group.py index 26c21121ad69f2..f6c295efe1dd32 100644 --- a/src/sentry/api/serializers/models/group.py +++ b/src/sentry/api/serializers/models/group.py @@ -52,19 +52,8 @@ from sentry.models.orgauthtoken import is_org_auth_token_auth from sentry.models.team import Team from sentry.models.user import User -from sentry.notifications.helpers import ( - collect_groups_by_project, - get_groups_for_query, - get_subscription_from_attributes, - get_user_subscriptions_for_groups, - should_use_notifications_v2, - transform_to_notification_settings_by_scope, -) -from sentry.notifications.types import ( - GroupSubscriptionStatus, - NotificationSettingEnum, - NotificationSettingTypes, -) +from sentry.notifications.helpers import collect_groups_by_project, get_subscription_from_attributes +from sentry.notifications.types import GroupSubscriptionStatus, NotificationSettingEnum from sentry.reprocessing2 import get_progress from sentry.search.events.constants import RELEASE_STAGE_ALIAS from sentry.search.events.filter import convert_search_filter_to_snuba_query, format_search_filter @@ -589,62 +578,39 @@ def _get_subscriptions( groups_by_project = collect_groups_by_project(groups) project_ids = list(groups_by_project.keys()) - if should_use_notifications_v2(groups[0].project.organization): - enabled_settings = notifications_service.get_subscriptions_for_projects( - user_id=user.id, project_ids=project_ids, type=NotificationSettingEnum.WORKFLOW - ) - query_groups = { - group for group in groups if (not enabled_settings[group.project_id][2]) - } - subscriptions_by_group_id: dict[int, GroupSubscription] = { - subscription.group_id: subscription - for subscription in GroupSubscription.objects.filter( - group__in=query_groups, user_id=user.id - ) - } - groups_by_project = collect_groups_by_project(groups) - - results = {} - for project_id, group_set in groups_by_project.items(): - s = enabled_settings[project_id] - subscription_status = GroupSubscriptionStatus( - is_disabled=s[0], - is_active=s[1], - has_only_inactive_subscriptions=s[2], - ) - for group in group_set: - subscription = subscriptions_by_group_id.get(group.id) - if subscription: - # Having a GroupSubscription overrides NotificationSettings. - results[group.id] = (False, subscription.is_active, subscription) - elif subscription_status.is_disabled: - # The user has disabled notifications in all cases. - results[group.id] = (True, False, None) - else: - # Since there is no subscription, it is only active if the value is ALWAYS. - results[group.id] = (False, subscription_status.is_active, None) - - return results - - notification_settings_by_scope = transform_to_notification_settings_by_scope( - notifications_service.get_settings_for_user_by_projects( - type=NotificationSettingTypes.WORKFLOW, - user_id=user.id, - parent_ids=project_ids, - ) + enabled_settings = notifications_service.get_subscriptions_for_projects( + user_id=user.id, project_ids=project_ids, type=NotificationSettingEnum.WORKFLOW ) - query_groups = get_groups_for_query(groups_by_project, notification_settings_by_scope, user) - subscriptions = GroupSubscription.objects.filter(group__in=query_groups, user_id=user.id) - subscriptions_by_group_id = { - subscription.group_id: subscription for subscription in subscriptions + query_groups = {group for group in groups if (not enabled_settings[group.project_id][2])} + subscriptions_by_group_id: dict[int, GroupSubscription] = { + subscription.group_id: subscription + for subscription in GroupSubscription.objects.filter( + group__in=query_groups, user_id=user.id + ) } + groups_by_project = collect_groups_by_project(groups) - return get_user_subscriptions_for_groups( - groups_by_project, - notification_settings_by_scope, - subscriptions_by_group_id, - user, - ) + results = {} + for project_id, group_set in groups_by_project.items(): + s = enabled_settings[project_id] + subscription_status = GroupSubscriptionStatus( + is_disabled=s[0], + is_active=s[1], + has_only_inactive_subscriptions=s[2], + ) + for group in group_set: + subscription = subscriptions_by_group_id.get(group.id) + if subscription: + # Having a GroupSubscription overrides NotificationSettings. + results[group.id] = (False, subscription.is_active, subscription) + elif subscription_status.is_disabled: + # The user has disabled notifications in all cases. + results[group.id] = (True, False, None) + else: + # Since there is no subscription, it is only active if the value is ALWAYS. + results[group.id] = (False, subscription_status.is_active, None) + + return results @staticmethod def _resolve_resolutions( @@ -766,7 +732,6 @@ def _is_authorized(user, organization_id: int): and getattr(request.user, "is_sentry_app", False) and is_api_token_auth(request.auth) ): - if AuthenticatedToken.from_token(request.auth).token_has_org_access(organization_id): return True diff --git a/src/sentry/conf/server.py b/src/sentry/conf/server.py index 2a6ab5bf39c136..d389d2b5179f80 100644 --- a/src/sentry/conf/server.py +++ b/src/sentry/conf/server.py @@ -1828,8 +1828,6 @@ def custom_parameter_sort(parameter: dict) -> tuple[str, int]: "organizations:sdk-crash-detection": False, # Enable functionality for recap server polling. "organizations:recap-server": False, - # Enable the new notification settings system - "organizations:notification-settings-v2": False, # Enable new release UI "organizations:release-ui-v2": False, # Enable User Feedback v2 ingest diff --git a/src/sentry/features/__init__.py b/src/sentry/features/__init__.py index a170a8fc1304c0..70f678a6a52377 100644 --- a/src/sentry/features/__init__.py +++ b/src/sentry/features/__init__.py @@ -262,7 +262,6 @@ default_manager.add("organizations:sourcemaps-bundle-flat-file-indexing", OrganizationFeature, FeatureHandlerStrategy.REMOTE) default_manager.add("organizations:sourcemaps-upload-release-as-artifact-bundle", OrganizationFeature, FeatureHandlerStrategy.REMOTE) default_manager.add("organizations:recap-server", OrganizationFeature, FeatureHandlerStrategy.INTERNAL) -default_manager.add("organizations:notification-settings-v2", OrganizationFeature, FeatureHandlerStrategy.INTERNAL) default_manager.add("organizations:release-ui-v2", OrganizationFeature, FeatureHandlerStrategy.REMOTE) default_manager.add("organizations:source-maps-debugger-blue-thunder-edition", OrganizationFeature, FeatureHandlerStrategy.REMOTE) default_manager.add("organizations:suspect-commits-all-frames", OrganizationFeature, FeatureHandlerStrategy.INTERNAL) diff --git a/src/sentry/incidents/action_handlers.py b/src/sentry/incidents/action_handlers.py index 607cc9686d896b..9250953b3a455c 100644 --- a/src/sentry/incidents/action_handlers.py +++ b/src/sentry/incidents/action_handlers.py @@ -20,10 +20,8 @@ IncidentStatus, TriggerStatus, ) -from sentry.models.notificationsetting import NotificationSetting from sentry.models.rulesnooze import RuleSnooze from sentry.models.user import User -from sentry.notifications.helpers import should_use_notifications_v2 from sentry.notifications.types import NotificationSettingEnum from sentry.notifications.utils.participants import get_notification_recipients_v2 from sentry.services.hybrid_cloud.actor import ActorType, RpcActor @@ -132,23 +130,17 @@ def _get_targets(self) -> Set[int]: elif self.action.target_type == AlertRuleTriggerAction.TargetType.TEAM.value: users = None - if should_use_notifications_v2(self.project.organization): - out = get_notification_recipients_v2( - recipients=list( - RpcActor(id=member.user_id, actor_type=ActorType.USER) - for member in target.member_set - ), - type=NotificationSettingEnum.ISSUE_ALERTS, - organization_id=self.project.organization_id, - project_ids=[self.project.id], - actor_type=ActorType.USER, - ) - users = out[ExternalProviders.EMAIL] - else: - users = NotificationSetting.objects.filter_to_accepting_recipients( - self.project, - {RpcUser(id=member.user_id) for member in target.member_set}, - )[ExternalProviders.EMAIL] + out = get_notification_recipients_v2( + recipients=list( + RpcActor(id=member.user_id, actor_type=ActorType.USER) + for member in target.member_set + ), + type=NotificationSettingEnum.ISSUE_ALERTS, + organization_id=self.project.organization_id, + project_ids=[self.project.id], + actor_type=ActorType.USER, + ) + users = out[ExternalProviders.EMAIL] snoozed_users = RuleSnooze.objects.filter( alert_rule=self.incident.alert_rule, user_id__in=[user.id for user in users] diff --git a/src/sentry/integrations/slack/views/link_identity.py b/src/sentry/integrations/slack/views/link_identity.py index 57d3297e89bd6c..a1693ac244061b 100644 --- a/src/sentry/integrations/slack/views/link_identity.py +++ b/src/sentry/integrations/slack/views/link_identity.py @@ -6,8 +6,6 @@ from sentry.integrations.utils import get_identity_or_404 from sentry.models.identity import Identity from sentry.models.integrations.integration import Integration -from sentry.models.notificationsetting import NotificationSetting -from sentry.notifications.helpers import should_use_notifications_v2 from sentry.notifications.notificationcontroller import NotificationController from sentry.notifications.notifications.integration_nudge import IntegrationNudgeNotification from sentry.types.integrations import ExternalProviderEnum, ExternalProviders @@ -71,19 +69,12 @@ def handle(self, request: Request, signed_params: str) -> HttpResponse: send_slack_response(integration, SUCCESS_LINKED_MESSAGE, params, command="link") has_slack_settings = None - if should_use_notifications_v2(organization): - controller = NotificationController( - recipients=[request.user], - organization_id=organization.id, - provider=ExternalProviderEnum.SLACK, - ) - has_slack_settings = controller.user_has_any_provider_settings( - ExternalProviderEnum.SLACK - ) - else: - has_slack_settings = NotificationSetting.objects.has_any_provider_settings( - request.user, ExternalProviders.SLACK - ) + controller = NotificationController( + recipients=[request.user], + organization_id=organization.id, + provider=ExternalProviderEnum.SLACK, + ) + has_slack_settings = controller.user_has_any_provider_settings(ExternalProviderEnum.SLACK) if not has_slack_settings: IntegrationNudgeNotification(organization, request.user, ExternalProviders.SLACK).send() diff --git a/src/sentry/mail/adapter.py b/src/sentry/mail/adapter.py index d78893b4e7f2ad..57df34ff920afd 100644 --- a/src/sentry/mail/adapter.py +++ b/src/sentry/mail/adapter.py @@ -6,10 +6,8 @@ from sentry.digests import Digest from sentry.digests import get_option_key as get_digest_option_key from sentry.digests.notifications import event_to_record, unsplit_key -from sentry.models.notificationsetting import NotificationSetting from sentry.models.options.project_option import ProjectOption from sentry.models.project import Project -from sentry.notifications.helpers import should_use_notifications_v2 from sentry.notifications.notificationcontroller import NotificationController from sentry.notifications.notifications.activity import EMAIL_CLASSES_BY_TYPE from sentry.notifications.notifications.digest import DigestNotification @@ -112,30 +110,18 @@ def get_sendable_user_objects(project): user_ids = project.member_set.values_list("user_id", flat=True) users = user_service.get_many(filter=dict(user_ids=list(user_ids))) - if should_use_notifications_v2(project.organization): - controller = NotificationController( - recipients=users, - project_ids=[project.id], - organization_id=project.organization_id, - provider=ExternalProviderEnum.EMAIL, - type=NotificationSettingEnum.ISSUE_ALERTS, - ) - return controller.get_notification_recipients( - type=NotificationSettingEnum.ISSUE_ALERTS, - actor_type=ActorType.USER, - )[ExternalProviders.EMAIL] - - accepting_recipients = NotificationSetting.objects.filter_to_accepting_recipients( - project, users + # TODO: Do we need to use a notification service here? + controller = NotificationController( + recipients=users, + project_ids=[project.id], + organization_id=project.organization_id, + provider=ExternalProviderEnum.EMAIL, + type=NotificationSettingEnum.ISSUE_ALERTS, ) - email_recipients = accepting_recipients.get(ExternalProviders.EMAIL, ()) - - users_by_id = {user.id: user for user in users} - return [ - users_by_id[recipient.id] - for recipient in email_recipients - if recipient.actor_type == ActorType.USER - ] + return controller.get_notification_recipients( + type=NotificationSettingEnum.ISSUE_ALERTS, + actor_type=ActorType.USER, + )[ExternalProviders.EMAIL] def get_sendable_user_ids(self, project): users = self.get_sendable_user_objects(project) diff --git a/src/sentry/models/groupsubscription.py b/src/sentry/models/groupsubscription.py index d7ebb504378d3b..f0a3fa6d4d6e6a 100644 --- a/src/sentry/models/groupsubscription.py +++ b/src/sentry/models/groupsubscription.py @@ -16,16 +16,10 @@ sane_repr, ) from sentry.db.models.fields.hybrid_cloud_foreign_key import HybridCloudForeignKey -from sentry.notifications.helpers import ( - should_use_notifications_v2, - transform_to_notification_settings_by_recipient, - where_should_be_participating, -) from sentry.notifications.types import ( GroupSubscriptionReason, NotificationSettingEnum, NotificationSettingsOptionEnum, - NotificationSettingTypes, ) from sentry.services.hybrid_cloud.actor import RpcActor from sentry.services.hybrid_cloud.notifications import notifications_service @@ -119,7 +113,6 @@ def bulk_subscribe( # 5 retries for race conditions where # concurrent subscription attempts cause integrity errors for i in range(4, -1, -1): # 4 3 2 1 0 - existing_subscriptions = set( GroupSubscription.objects.filter( user_id__in=user_ids, group=group, project=group.project @@ -186,81 +179,53 @@ def get_participants(self, group: Group) -> ParticipantMap: "organizations:team-workflow-notifications", group.project.organization ) - if should_use_notifications_v2(group.project.organization) and has_team_workflow: + if has_team_workflow: possible_team_actors = self.get_possible_team_actors(group) all_possible_actors += possible_team_actors subscriptions_by_team_id = self.get_subscriptions_by_team_id( group, possible_team_actors ) - if should_use_notifications_v2(group.project.organization): - if not all_possible_actors: # no actors, no notifications - return ParticipantMap() + if not all_possible_actors: # no actors, no notifications + return ParticipantMap() - providers_by_recipient = notifications_service.get_participants( - recipients=all_possible_actors, - project_ids=[group.project_id], - organization_id=group.organization.id, - type=NotificationSettingEnum.WORKFLOW, - ) - result = ParticipantMap() - for user in all_possible_actors: - if user.id not in providers_by_recipient: - continue - - subscription_option = subscriptions_by_user_id.get(user.id, {}) - if not subscription_option and has_team_workflow: - subscription_option = subscriptions_by_team_id.get(user.id, {}) - - for provider_str, val in providers_by_recipient[user.id].items(): - value = NotificationSettingsOptionEnum(val) - is_subscribed = ( - subscription_option - and subscription_option.is_active - and value - in [ - NotificationSettingsOptionEnum.ALWAYS, - NotificationSettingsOptionEnum.SUBSCRIBE_ONLY, - ] - ) - is_implicit = ( - not subscription_option and value == NotificationSettingsOptionEnum.ALWAYS - ) - if is_subscribed or is_implicit: - reason = ( - subscription_option - and subscription_option.reason - or GroupSubscriptionReason.implicit - ) - provider = ExternalProviders(provider_str) - result.add(provider, user, reason) - return result - - notification_settings = notifications_service.get_settings_for_recipient_by_parent( - type=NotificationSettingTypes.WORKFLOW, + providers_by_recipient = notifications_service.get_participants( recipients=all_possible_actors, - parent_id=group.project_id, - ) - notification_settings_by_recipient = transform_to_notification_settings_by_recipient( - notification_settings, all_possible_actors + project_ids=[group.project_id], + organization_id=group.organization.id, + type=NotificationSettingEnum.WORKFLOW, ) - result = ParticipantMap() for user in all_possible_actors: - subscription_option = subscriptions_by_user_id.get(user.id) - providers = where_should_be_participating( - user, - subscription_option, - notification_settings_by_recipient, - ) - for provider in providers: - reason = ( + if user.id not in providers_by_recipient: + continue + + subscription_option = subscriptions_by_user_id.get(user.id, {}) + if not subscription_option and has_team_workflow: + subscription_option = subscriptions_by_team_id.get(user.id, {}) + + for provider_str, val in providers_by_recipient[user.id].items(): + value = NotificationSettingsOptionEnum(val) + is_subscribed = ( subscription_option - and subscription_option.reason - or GroupSubscriptionReason.implicit + and subscription_option.is_active + and value + in [ + NotificationSettingsOptionEnum.ALWAYS, + NotificationSettingsOptionEnum.SUBSCRIBE_ONLY, + ] ) - result.add(provider, user, reason) - + is_implicit = ( + not subscription_option and value == NotificationSettingsOptionEnum.ALWAYS + ) + if is_subscribed or is_implicit: + reason = ( + subscription_option + and subscription_option.reason + or GroupSubscriptionReason.implicit + ) + provider = ExternalProviders(provider_str) + result.add(provider, user, reason) return result def get_possible_team_actors(self, group: Group) -> List[RpcActor]: diff --git a/src/sentry/notifications/manager.py b/src/sentry/notifications/manager.py index 516d6697d4975c..d6e8ac2a2a662d 100644 --- a/src/sentry/notifications/manager.py +++ b/src/sentry/notifications/manager.py @@ -25,14 +25,7 @@ from sentry.models.team import Team from sentry.models.user import User from sentry.notifications.defaults import NOTIFICATION_SETTINGS_ALL_SOMETIMES -from sentry.notifications.helpers import ( - get_scope, - get_scope_type, - should_use_notifications_v2, - transform_to_notification_settings_by_recipient, - validate, - where_should_recipient_be_notified, -) +from sentry.notifications.helpers import get_scope, get_scope_type, validate from sentry.notifications.notificationcontroller import NotificationController from sentry.notifications.types import ( NOTIFICATION_SCOPE_TYPE, @@ -48,7 +41,6 @@ NotificationSettingTypes, ) from sentry.services.hybrid_cloud.actor import ActorType, RpcActor -from sentry.services.hybrid_cloud.notifications import notifications_service from sentry.services.hybrid_cloud.user.model import RpcUser from sentry.types.integrations import ( EXTERNAL_PROVIDERS, @@ -424,40 +416,20 @@ def filter_to_accepting_recipients( organization = parent project_ids = None - if should_use_notifications_v2(organization): - # We should replace calls to NotificationSettings.get_notification_recipients at the call site - this code should never be reached - setting_type = ( - NotificationSettingEnum(NOTIFICATION_SETTING_TYPES[type]) - if type - else NotificationSettingEnum.ISSUE_ALERTS - ) - controller = NotificationController( - recipients=recipient_actors, - project_ids=project_ids, - organization_id=organization.id, - type=setting_type, - ) - - logger.warning("Missing upstream implementation for get_notification_recipients in v2") - return controller.get_notification_recipients(type=setting_type) - - notification_settings = notifications_service.get_settings_for_recipient_by_parent( - type=type, parent_id=parent.id, recipients=recipient_actors + setting_type = ( + NotificationSettingEnum(NOTIFICATION_SETTING_TYPES[type]) + if type + else NotificationSettingEnum.ISSUE_ALERTS ) - notification_settings_by_recipient = transform_to_notification_settings_by_recipient( - notification_settings, recipient_actors + controller = NotificationController( + recipients=recipient_actors, + project_ids=project_ids, + organization_id=organization.id, + type=setting_type, ) - mapping = defaultdict(set) - for recipient in recipient_actors: - providers = where_should_recipient_be_notified( - notification_settings_by_recipient, recipient, type - ) - for provider in providers: - mapping[provider].add(recipient) - return mapping + return controller.get_notification_recipients(type=setting_type) - # TODO(snigdha): cleanup after v2 def get_notification_recipients( self, project: Project ) -> Mapping[ExternalProviders, Iterable[RpcActor]]: @@ -575,7 +547,7 @@ def update_settings_bulk( id_key = "team_id" id = team.id - for (provider, type, scope_type, scope_identifier, value) in notification_settings: + for provider, type, scope_type, scope_identifier, value in notification_settings: # A missing DB row is equivalent to DEFAULT. if value == NotificationSettingOptionValues.DEFAULT: self._filter( @@ -614,7 +586,7 @@ def update_settings_bulk( enabled_value_dict = {} # group the type, scope_type, scope_identifier, together and get store the explicitly enabled/disabled providers - for (provider, type, scope_type, scope_identifier, value) in notification_settings: + for provider, type, scope_type, scope_identifier, value in notification_settings: # Group the type, scope_type, scope_identifier together group_key = (type, scope_type, scope_identifier) all_settings.add(group_key) diff --git a/src/sentry/notifications/notifications/activity/new_processing_issues.py b/src/sentry/notifications/notifications/activity/new_processing_issues.py index 6b416e2cc37109..b84a9ce5106349 100644 --- a/src/sentry/notifications/notifications/activity/new_processing_issues.py +++ b/src/sentry/notifications/notifications/activity/new_processing_issues.py @@ -4,8 +4,6 @@ from urllib.parse import urlencode from sentry.models.activity import Activity -from sentry.models.notificationsetting import NotificationSetting -from sentry.notifications.helpers import should_use_notifications_v2 from sentry.notifications.notificationcontroller import NotificationController from sentry.notifications.types import GroupSubscriptionReason, NotificationSettingEnum from sentry.notifications.utils import summarize_issues @@ -27,21 +25,17 @@ def __init__(self, activity: Activity) -> None: def get_participants_with_group_subscription_reason(self) -> ParticipantMap: participants_by_provider = None - if should_use_notifications_v2(self.project.organization): - user_ids = list(self.project.member_set.values_list("user_id", flat=True)) - users = user_service.get_many(filter={"user_ids": user_ids}) - notification_controller = NotificationController( - recipients=users, - project_ids=[self.project.id], - organization_id=self.project.organization_id, - ) - participants_by_provider = notification_controller.get_notification_recipients( - type=NotificationSettingEnum.WORKFLOW, - ) - else: - participants_by_provider = NotificationSetting.objects.get_notification_recipients( - self.project - ) + user_ids = list(self.project.member_set.values_list("user_id", flat=True)) + users = user_service.get_many(filter={"user_ids": user_ids}) + # TODO: Do we need to use a notification service here? + notification_controller = NotificationController( + recipients=users, + project_ids=[self.project.id], + organization_id=self.project.organization_id, + ) + participants_by_provider = notification_controller.get_notification_recipients( + type=NotificationSettingEnum.WORKFLOW, + ) result = ParticipantMap() for provider, participants in participants_by_provider.items(): diff --git a/src/sentry/notifications/notifications/base.py b/src/sentry/notifications/notifications/base.py index 8c725df5bb9465..5e8774783c1e4a 100644 --- a/src/sentry/notifications/notifications/base.py +++ b/src/sentry/notifications/notifications/base.py @@ -10,8 +10,6 @@ from sentry import analytics from sentry.db.models import Model from sentry.models.environment import Environment -from sentry.models.notificationsetting import NotificationSetting -from sentry.notifications.helpers import should_use_notifications_v2 from sentry.notifications.types import ( NOTIFICATION_SETTING_TYPES, NotificationSettingEnum, @@ -240,21 +238,11 @@ def filter_to_accepting_recipients( if self.notification_setting_type else NotificationSettingEnum.ISSUE_ALERTS ) - if should_use_notifications_v2(self.organization): - return get_notification_recipients_v2( - recipients=recipients, - type=setting_type, - organization_id=self.organization.id, - ) - - accepting_recipients: Mapping[ - ExternalProviders, Iterable[RpcActor] - ] = NotificationSetting.objects.filter_to_accepting_recipients( - self.organization, - recipients, - self.notification_setting_type or NotificationSettingTypes.ISSUE_ALERTS, + return get_notification_recipients_v2( + recipients=recipients, + type=setting_type, + organization_id=self.organization.id, ) - return accepting_recipients def get_participants(self) -> Mapping[ExternalProviders, Iterable[RpcActor]]: # need a notification_setting_type to call this function diff --git a/src/sentry/notifications/utils/participants.py b/src/sentry/notifications/utils/participants.py index 4502c56ceebf7b..2cbaff5a04ad85 100644 --- a/src/sentry/notifications/utils/participants.py +++ b/src/sentry/notifications/utils/participants.py @@ -23,7 +23,6 @@ from sentry.models.group import Group from sentry.models.groupassignee import GroupAssignee from sentry.models.groupsubscription import GroupSubscription -from sentry.models.notificationsetting import NotificationSetting from sentry.models.organization import Organization from sentry.models.organizationmember import OrganizationMember from sentry.models.organizationmemberteam import OrganizationMemberTeam @@ -34,12 +33,6 @@ from sentry.models.rulesnooze import RuleSnooze from sentry.models.team import Team from sentry.models.user import User -from sentry.notifications.helpers import ( - get_values_by_provider_by_type, - should_use_notifications_v2, - transform_to_notification_settings_by_recipient, -) -from sentry.notifications.notify import notification_providers from sentry.notifications.types import ( NOTIFICATION_SETTING_TYPES, ActionTargetType, @@ -92,7 +85,7 @@ def add_all(self, provider: ExternalProviders, actor_group: Mapping[RpcActor, in self._dict[provider].update(actor_group) def update(self, other: ParticipantMap) -> None: - for (provider, actor_group) in other._dict.items(): + for provider, actor_group in other._dict.items(): self.add_all(provider, actor_group) def get_participant_sets(self) -> Iterable[Tuple[ExternalProviders, Iterable[RpcActor]]]: @@ -213,51 +206,22 @@ def get_participants_for_release( ) actors = RpcActor.many_from_object(RpcUser(id=user_id) for user_id in user_ids) - if should_use_notifications_v2(organization): - # don't pass in projects since the settings are scoped to the organization only for now - providers_by_recipient = notifications_service.get_participants( - type=NotificationSettingEnum.DEPLOY, - recipients=actors, - organization_id=organization.id, - ) - - users_to_reasons_by_provider = ParticipantMap() - for actor in actors: - settings = providers_by_recipient.get(actor.id, {}) - for provider_str, val_str in settings.items(): - provider = ExternalProviders(provider_str) - val = NotificationSettingsOptionEnum(val_str) - reason = get_reason(actor, val, commited_user_ids) - if reason: - users_to_reasons_by_provider.add(provider, actor, reason) - return users_to_reasons_by_provider - - # Get all the involved users' settings for deploy-emails (including - # users' organization-independent settings.) - notification_settings = notifications_service.get_settings_for_recipient_by_parent( - type=NotificationSettingTypes.DEPLOY, + # don't pass in projects since the settings are scoped to the organization only for now + providers_by_recipient = notifications_service.get_participants( + type=NotificationSettingEnum.DEPLOY, recipients=actors, - parent_id=organization.id, - ) - notification_settings_by_recipient = transform_to_notification_settings_by_recipient( - notification_settings, actors + organization_id=organization.id, ) - # Map users to their setting value. Prioritize user/org specific, then - # user default, then product default. users_to_reasons_by_provider = ParticipantMap() for actor in actors: - notification_settings_by_scope = notification_settings_by_recipient.get(actor, {}) - values_by_provider = get_values_by_provider_by_type( - notification_settings_by_scope, - notification_providers(), - NotificationSettingTypes.DEPLOY, - actor, - ) - for provider, value in values_by_provider.items(): - reason_option = get_reason(actor, value, commited_user_ids) - if reason_option: - users_to_reasons_by_provider.add(provider, actor, reason_option) + settings = providers_by_recipient.get(actor.id, {}) + for provider_str, val_str in settings.items(): + provider = ExternalProviders(provider_str) + val = NotificationSettingsOptionEnum(val_str) + reason = get_reason(actor, val, commited_user_ids) + if reason: + users_to_reasons_by_provider.add(provider, actor, reason) return users_to_reasons_by_provider @@ -606,19 +570,14 @@ def get_recipients_by_provider( setting_type = NotificationSettingEnum(NOTIFICATION_SETTING_TYPES[notification_type]) teams_by_provider: Mapping[ExternalProviders, Iterable[RpcActor]] = {} - if should_use_notifications_v2(project.organization): - # get by team - teams_by_provider = get_notification_recipients_v2( - recipients=teams, - type=setting_type, - organization_id=project.organization_id, - project_ids=[project.id], - actor_type=ActorType.TEAM, - ) - else: - teams_by_provider = NotificationSetting.objects.filter_to_accepting_recipients( - project, teams, notification_type - ) + # get by team + teams_by_provider = get_notification_recipients_v2( + recipients=teams, + type=setting_type, + organization_id=project.organization_id, + project_ids=[project.id], + actor_type=ActorType.TEAM, + ) # Teams cannot receive emails so omit EMAIL settings. teams_by_provider = { @@ -632,18 +591,13 @@ def get_recipients_by_provider( # Repeat for users. users_by_provider: Mapping[ExternalProviders, Iterable[RpcActor]] = {} - if should_use_notifications_v2(project.organization): - # convert from string to enum - users_by_provider = get_notification_recipients_v2( - recipients=users, - type=setting_type, - organization_id=project.organization_id, - project_ids=[project.id], - actor_type=ActorType.USER, - ) - else: - users_by_provider = NotificationSetting.objects.filter_to_accepting_recipients( - project, users, notification_type - ) + # convert from string to enum + users_by_provider = get_notification_recipients_v2( + recipients=users, + type=setting_type, + organization_id=project.organization_id, + project_ids=[project.id], + actor_type=ActorType.USER, + ) return combine_recipients_by_provider(teams_by_provider, users_by_provider) diff --git a/src/sentry/plugins/bases/notify.py b/src/sentry/plugins/bases/notify.py index 72ceef30f57fde..0005180f895766 100644 --- a/src/sentry/plugins/bases/notify.py +++ b/src/sentry/plugins/bases/notify.py @@ -8,8 +8,6 @@ from sentry import digests, ratelimits from sentry.exceptions import InvalidIdentity, PluginError -from sentry.models.notificationsetting import NotificationSetting -from sentry.notifications.helpers import should_use_notifications_v2 from sentry.notifications.notificationcontroller import NotificationController from sentry.notifications.types import NotificationSettingEnum from sentry.plugins.base import Notification, Plugin @@ -146,23 +144,19 @@ def get_sendable_user_objects(self, project): notifications for the provided project. """ if self.get_conf_key() == "mail": - if should_use_notifications_v2(self.project.organization): - user_ids = list(project.member_set.values_list("user_id", flat=True)) - users = user_service.get_many(filter={"user_ids": user_ids}) - notification_controller = NotificationController( - recipients=users, - project_ids=[project.id], - organization_id=project.organization_id, - provider=ExternalProviderEnum.EMAIL, - ) - return notification_controller.get_notification_recipients( - type=NotificationSettingEnum.ISSUE_ALERTS, - actor_type=ActorType.USER, - )[ExternalProviders.EMAIL] - else: - return NotificationSetting.objects.get_notification_recipients(project)[ - ExternalProviders.EMAIL - ] + user_ids = list(project.member_set.values_list("user_id", flat=True)) + users = user_service.get_many(filter={"user_ids": user_ids}) + # TODO: Do we need to use a notification service here? + notification_controller = NotificationController( + recipients=users, + project_ids=[project.id], + organization_id=project.organization_id, + provider=ExternalProviderEnum.EMAIL, + ) + return notification_controller.get_notification_recipients( + type=NotificationSettingEnum.ISSUE_ALERTS, + actor_type=ActorType.USER, + )[ExternalProviders.EMAIL] return self.get_notification_recipients(project, f"{self.get_conf_key()}:alert") diff --git a/src/sentry/tasks/weekly_reports.py b/src/sentry/tasks/weekly_reports.py index 5ae0103ef761c6..1a447fa30b5ae0 100644 --- a/src/sentry/tasks/weekly_reports.py +++ b/src/sentry/tasks/weekly_reports.py @@ -28,9 +28,7 @@ from sentry.models.grouphistory import GroupHistory, GroupHistoryStatus from sentry.models.organization import Organization, OrganizationStatus from sentry.models.organizationmember import OrganizationMember -from sentry.notifications.helpers import should_use_notifications_v2 from sentry.services.hybrid_cloud.notifications import notifications_service -from sentry.services.hybrid_cloud.user_option import user_option_service from sentry.silo import SiloMode from sentry.snuba.dataset import Dataset from sentry.tasks.base import instrumented_task, retry @@ -222,25 +220,20 @@ def prepare_organization_report( ) return - use_notifications_v2 = should_use_notifications_v2(ctx.organization) - # Finally, deliver the reports with sentry_sdk.start_span(op="weekly_reports.deliver_reports"): deliver_reports( - ctx, - dry_run=dry_run, - target_user=target_user, - email_override=email_override, - use_notifications_v2=use_notifications_v2, + ctx, dry_run=dry_run, target_user=target_user, email_override=email_override ) # Organization Passes + # Find the projects associated with an user. # Populates context.project_ownership which is { user_id: set<project_id> } def user_project_ownership(ctx): - for (project_id, user_id) in OrganizationMember.objects.filter( + for project_id, user_id in OrganizationMember.objects.filter( organization_id=ctx.organization.id, teams__projectteam__project__isnull=False ).values_list("teams__projectteam__project_id", "user_id"): ctx.project_ownership.setdefault(user_id, set()).add(project_id) @@ -651,13 +644,11 @@ def fetch_key_performance_issue_groups(ctx): # For all users in the organization, we generate the template context for the user, and send the email. -def deliver_reports( - ctx, dry_run=False, target_user=None, email_override=None, use_notifications_v2=False -): +def deliver_reports(ctx, dry_run=False, target_user=None, email_override=None): # Specify a sentry user to send this email. if email_override: send_email(ctx, target_user, dry_run=dry_run, email_override=email_override) - elif use_notifications_v2: + else: user_list = list( OrganizationMember.objects.filter( user_is_active=True, @@ -673,30 +664,6 @@ def deliver_reports( for user_id in user_ids: send_email(ctx, user_id, dry_run=dry_run) - else: - # We save the subscription status of the user in a field in UserOptions. - user_list = list( - OrganizationMember.objects.filter( - user_is_active=True, - organization_id=ctx.organization.id, - ) - .filter(flags=F("flags").bitand(~OrganizationMember.flags["member-limit:restricted"])) - .values_list("user_id", flat=True) - ) - user_list = list(filter(lambda v: v is not None, user_list)) - options_by_user_id = { - option.user_id: option.value - for option in user_option_service.get_many( - filter=dict(user_ids=user_list, keys=["reports:disabled-organizations"]) - ) - } - - for user_id in user_list: - option = list(options_by_user_id.get(user_id, [])) - user_subscribed_to_organization_reports = ctx.organization.id not in option - if user_subscribed_to_organization_reports: - send_email(ctx, user_id, dry_run=dry_run) - project_breakdown_colors = ["#422C6E", "#895289", "#D6567F", "#F38150", "#F2B713"] total_color = """ @@ -1022,7 +989,7 @@ def all_key_transactions(): def key_performance_issues(): def all_key_performance_issues(): for project_ctx in user_projects: - for (group, group_history, count) in project_ctx.key_performance_issues: + for group, group_history, count in project_ctx.key_performance_issues: yield { "count": count, "group": group, diff --git a/tests/sentry/tasks/test_weekly_reports.py b/tests/sentry/tasks/test_weekly_reports.py index 79fa3181c363c0..725069e8e923d9 100644 --- a/tests/sentry/tasks/test_weekly_reports.py +++ b/tests/sentry/tasks/test_weekly_reports.py @@ -1,5 +1,4 @@ import copy -import functools from datetime import datetime, timedelta, timezone from unittest import mock @@ -12,7 +11,7 @@ from sentry.constants import DataCategory from sentry.models.group import GroupStatus from sentry.models.grouphistory import GroupHistoryStatus -from sentry.models.options.user_option import UserOption +from sentry.models.notificationsettingoption import NotificationSettingOption from sentry.models.organizationmember import OrganizationMember from sentry.models.project import Project from sentry.services.hybrid_cloud.user_option import user_option_service @@ -127,19 +126,23 @@ def test_deliver_reports_respects_settings(self, mock_send_email): organization = self.organization ctx = OrganizationReportContext(0, 0, organization) - set_option_value = assume_test_silo_mode(SiloMode.CONTROL)( - functools.partial( - UserOption.objects.set_value, user, DISABLED_ORGANIZATIONS_USER_OPTION_KEY - ) - ) + def set_option_value(value): + with assume_test_silo_mode(SiloMode.CONTROL): + NotificationSettingOption.objects.update_or_create( + scope_type="organization", + scope_identifier=organization.id, + user_id=user.id, + type="reports", + defaults={"value": value}, + ) # disabled - set_option_value([organization.id]) + set_option_value("never") deliver_reports(ctx) assert mock_send_email.call_count == 0 # enabled - set_option_value([]) + set_option_value("always") deliver_reports(ctx) mock_send_email.assert_called_once_with(ctx, user.id, dry_run=False) @@ -176,7 +179,7 @@ def test_invited_member(self, mock_send_email): organization=self.organization, email="[email protected]", token="abc" ) - deliver_reports(ctx, use_notifications_v2=True) + deliver_reports(ctx) assert mock_send_email.call_count == 1 def test_organization_project_issue_summaries(self): @@ -468,7 +471,6 @@ def test_message_builder_substatus_simple(self, message_builder): @mock.patch("sentry.tasks.weekly_reports.MessageBuilder") def test_message_builder_advanced(self, message_builder): - now = django_timezone.now() two_days_ago = now - timedelta(days=2) three_days_ago = now - timedelta(days=3) @@ -562,7 +564,6 @@ def test_empty_report(self, mock_send_email): @with_feature("organizations:session-replay-weekly_report") @mock.patch("sentry.tasks.weekly_reports.MessageBuilder") def test_message_builder_replays(self, message_builder): - now = django_timezone.now() two_days_ago = now - timedelta(days=2) timestamp = to_timestamp(floor_to_utc_day(now))
35c39f7f44e0306f9ab431dfd1661423bf234f69
2024-09-28 00:50:15
Christinarlong
chore(sentry_apps): Move PlatformExternalIssue model to sentry_apps (#78076)
false
Move PlatformExternalIssue model to sentry_apps (#78076)
chore
diff --git a/src/sentry/api/endpoints/group_external_issue_details.py b/src/sentry/api/endpoints/group_external_issue_details.py index 13a730a620fcdd..1e04086f983198 100644 --- a/src/sentry/api/endpoints/group_external_issue_details.py +++ b/src/sentry/api/endpoints/group_external_issue_details.py @@ -5,7 +5,7 @@ from sentry.api.api_publish_status import ApiPublishStatus from sentry.api.base import region_silo_endpoint from sentry.api.bases.group import GroupEndpoint -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue @region_silo_endpoint diff --git a/src/sentry/api/endpoints/group_external_issues.py b/src/sentry/api/endpoints/group_external_issues.py index b1c812c432d81f..71ecceb05b00b4 100644 --- a/src/sentry/api/endpoints/group_external_issues.py +++ b/src/sentry/api/endpoints/group_external_issues.py @@ -5,7 +5,7 @@ from sentry.api.base import region_silo_endpoint from sentry.api.bases.group import GroupEndpoint from sentry.api.serializers import serialize -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue @region_silo_endpoint diff --git a/src/sentry/api/serializers/models/group.py b/src/sentry/api/serializers/models/group.py index 924aafb1d71e56..e056c3113faf91 100644 --- a/src/sentry/api/serializers/models/group.py +++ b/src/sentry/api/serializers/models/group.py @@ -648,7 +648,7 @@ def _resolve_resolutions( @staticmethod def _resolve_external_issue_annotations(groups: Sequence[Group]) -> Mapping[int, Sequence[Any]]: - from sentry.models.platformexternalissue import PlatformExternalIssue + from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue # find the external issues for sentry apps and add them in return ( diff --git a/src/sentry/api/serializers/models/group_stream.py b/src/sentry/api/serializers/models/group_stream.py index a77b8287592baf..78eaf1e48262dd 100644 --- a/src/sentry/api/serializers/models/group_stream.py +++ b/src/sentry/api/serializers/models/group_stream.py @@ -30,10 +30,10 @@ from sentry.models.groupinbox import get_inbox_details from sentry.models.grouplink import GroupLink from sentry.models.groupowner import get_owner_details -from sentry.models.platformexternalissue import PlatformExternalIssue from sentry.sentry_apps.api.serializers.platform_external_issue import ( PlatformExternalIssueSerializer, ) +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.snuba.dataset import Dataset from sentry.tsdb.base import TSDBModel from sentry.utils import metrics diff --git a/src/sentry/deletions/defaults/platform_external_issue.py b/src/sentry/deletions/defaults/platform_external_issue.py index ac8ecc31328299..d17e208dfb2dbc 100644 --- a/src/sentry/deletions/defaults/platform_external_issue.py +++ b/src/sentry/deletions/defaults/platform_external_issue.py @@ -1,7 +1,7 @@ from collections.abc import Sequence from sentry.deletions.base import ModelDeletionTask -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue class PlatformExternalIssueDeletionTask(ModelDeletionTask[PlatformExternalIssue]): diff --git a/src/sentry/issues/endpoints/group_details.py b/src/sentry/issues/endpoints/group_details.py index 7ae45cdf2501ef..a4ca1e2f19ab83 100644 --- a/src/sentry/issues/endpoints/group_details.py +++ b/src/sentry/issues/endpoints/group_details.py @@ -37,13 +37,13 @@ from sentry.models.groupowner import get_owner_details from sentry.models.groupseen import GroupSeen from sentry.models.groupsubscription import GroupSubscriptionManager -from sentry.models.platformexternalissue import PlatformExternalIssue from sentry.models.team import Team from sentry.models.userreport import UserReport from sentry.plugins.base import plugins from sentry.sentry_apps.api.serializers.platform_external_issue import ( PlatformExternalIssueSerializer, ) +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.tasks.post_process import fetch_buffered_group_stats from sentry.types.ratelimit import RateLimit, RateLimitCategory from sentry.users.services.user.service import user_service diff --git a/src/sentry/mediators/external_issues/creator.py b/src/sentry/mediators/external_issues/creator.py index 8ab273bcbc3527..f436b378f36bda 100644 --- a/src/sentry/mediators/external_issues/creator.py +++ b/src/sentry/mediators/external_issues/creator.py @@ -5,7 +5,7 @@ from sentry.mediators.mediator import Mediator from sentry.mediators.param import Param from sentry.models.group import Group -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.sentry_apps.services.app import RpcSentryAppInstallation diff --git a/src/sentry/mediators/external_issues/issue_link_creator.py b/src/sentry/mediators/external_issues/issue_link_creator.py index ac23160fc7e813..0228a23c545e18 100644 --- a/src/sentry/mediators/external_issues/issue_link_creator.py +++ b/src/sentry/mediators/external_issues/issue_link_creator.py @@ -7,7 +7,7 @@ from sentry.mediators.mediator import Mediator from sentry.mediators.param import Param from sentry.models.group import Group -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.sentry_apps.services.app import RpcSentryAppInstallation from sentry.users.services.user import RpcUser diff --git a/src/sentry/models/__init__.py b/src/sentry/models/__init__.py index f300d344c9a829..563978898e92b3 100644 --- a/src/sentry/models/__init__.py +++ b/src/sentry/models/__init__.py @@ -1,3 +1,4 @@ +from ..sentry_apps.models.platformexternalissue import * # NOQA from .activity import * # NOQA from .apiapplication import * # NOQA from .apiauthorization import * # NOQA @@ -75,7 +76,6 @@ from .organizationslugreservation import * # NOQA from .organizationslugreservationreplica import * # NOQA from .orgauthtoken import * # NOQA -from .platformexternalissue import * # NOQA from .project import * # NOQA from .projectbookmark import * # NOQA from .projectcodeowners import * # NOQA diff --git a/src/sentry/search/snuba/backend.py b/src/sentry/search/snuba/backend.py index 92aa4681f0623c..de47f44053ad3f 100644 --- a/src/sentry/search/snuba/backend.py +++ b/src/sentry/search/snuba/backend.py @@ -27,7 +27,6 @@ from sentry.models.grouplink import GroupLink from sentry.models.groupowner import GroupOwner from sentry.models.groupsubscription import GroupSubscription -from sentry.models.platformexternalissue import PlatformExternalIssue from sentry.models.project import Project from sentry.models.release import Release from sentry.models.team import Team @@ -40,6 +39,7 @@ PostgresSnubaQueryExecutor, TrendsSortWeights, ) +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.users.models.user import User from sentry.utils import metrics from sentry.utils.cursors import Cursor, CursorResult diff --git a/src/sentry/sentry_apps/api/endpoints/installation_external_issue_details.py b/src/sentry/sentry_apps/api/endpoints/installation_external_issue_details.py index 3197a29ad231d6..4cd3a5c820c689 100644 --- a/src/sentry/sentry_apps/api/endpoints/installation_external_issue_details.py +++ b/src/sentry/sentry_apps/api/endpoints/installation_external_issue_details.py @@ -7,7 +7,7 @@ from sentry.api.bases import ( SentryAppInstallationExternalIssueBaseEndpoint as ExternalIssueBaseEndpoint, ) -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue @region_silo_endpoint diff --git a/src/sentry/sentry_apps/api/serializers/platform_external_issue.py b/src/sentry/sentry_apps/api/serializers/platform_external_issue.py index 109a4cc9002520..e0c2398b78e088 100644 --- a/src/sentry/sentry_apps/api/serializers/platform_external_issue.py +++ b/src/sentry/sentry_apps/api/serializers/platform_external_issue.py @@ -1,5 +1,5 @@ from sentry.api.serializers import Serializer, register -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue @register(PlatformExternalIssue) diff --git a/src/sentry/sentry_apps/models/__init__.py b/src/sentry/sentry_apps/models/__init__.py index 05288934069583..4df7950fb9e52c 100644 --- a/src/sentry/sentry_apps/models/__init__.py +++ b/src/sentry/sentry_apps/models/__init__.py @@ -1,3 +1,4 @@ +from .platformexternalissue import PlatformExternalIssue from .sentry_app import SentryApp from .sentry_app_component import SentryAppComponent from .sentry_app_installation import SentryAppInstallation @@ -12,4 +13,5 @@ "ServiceHook", "SentryAppInstallationForProvider", "SentryAppComponent", + "PlatformExternalIssue", ) diff --git a/src/sentry/models/platformexternalissue.py b/src/sentry/sentry_apps/models/platformexternalissue.py similarity index 100% rename from src/sentry/models/platformexternalissue.py rename to src/sentry/sentry_apps/models/platformexternalissue.py diff --git a/src/sentry/testutils/factories.py b/src/sentry/testutils/factories.py index 4800df27028631..8116ac65529ed5 100644 --- a/src/sentry/testutils/factories.py +++ b/src/sentry/testutils/factories.py @@ -114,7 +114,6 @@ from sentry.models.organizationmemberteam import OrganizationMemberTeam from sentry.models.organizationslugreservation import OrganizationSlugReservation from sentry.models.orgauthtoken import OrgAuthToken -from sentry.models.platformexternalissue import PlatformExternalIssue from sentry.models.project import Project from sentry.models.projectbookmark import ProjectBookmark from sentry.models.projectcodeowners import ProjectCodeOwners @@ -136,6 +135,7 @@ SentryAppInstallationTokenCreator, ) from sentry.sentry_apps.logic import SentryAppCreator +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.sentry_apps.models.sentry_app import SentryApp from sentry.sentry_apps.models.sentry_app_installation import SentryAppInstallation from sentry.sentry_apps.models.sentry_app_installation_for_provider import ( diff --git a/tests/sentry/api/endpoints/test_group_external_issue_details.py b/tests/sentry/api/endpoints/test_group_external_issue_details.py index b8ce7151d93424..da7299a5ed54f6 100644 --- a/tests/sentry/api/endpoints/test_group_external_issue_details.py +++ b/tests/sentry/api/endpoints/test_group_external_issue_details.py @@ -1,4 +1,4 @@ -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.testutils.cases import APITestCase diff --git a/tests/sentry/api/endpoints/test_group_external_issues.py b/tests/sentry/api/endpoints/test_group_external_issues.py index 8a8fdff1e678a8..e6b1028058c305 100644 --- a/tests/sentry/api/endpoints/test_group_external_issues.py +++ b/tests/sentry/api/endpoints/test_group_external_issues.py @@ -1,4 +1,4 @@ -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.testutils.cases import APITestCase diff --git a/tests/sentry/issues/endpoints/test_organization_group_index.py b/tests/sentry/issues/endpoints/test_organization_group_index.py index 78e75243e2d48a..74b3153878c78a 100644 --- a/tests/sentry/issues/endpoints/test_organization_group_index.py +++ b/tests/sentry/issues/endpoints/test_organization_group_index.py @@ -45,7 +45,6 @@ from sentry.models.groupsnooze import GroupSnooze from sentry.models.groupsubscription import GroupSubscription from sentry.models.grouptombstone import GroupTombstone -from sentry.models.platformexternalissue import PlatformExternalIssue from sentry.models.release import Release from sentry.models.releaseprojectenvironment import ReleaseStages from sentry.models.savedsearch import SavedSearch, Visibility @@ -56,6 +55,7 @@ SEMVER_PACKAGE_ALIAS, ) from sentry.search.snuba.executors import GroupAttributesPostgresSnubaQueryExecutor +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.silo.base import SiloMode from sentry.testutils.cases import APITestCase, SnubaTestCase from sentry.testutils.helpers import parse_link_header diff --git a/tests/sentry/mediators/external_issues/test_creator.py b/tests/sentry/mediators/external_issues/test_creator.py index 3cab76d1d10f72..661cd256667a1f 100644 --- a/tests/sentry/mediators/external_issues/test_creator.py +++ b/tests/sentry/mediators/external_issues/test_creator.py @@ -1,5 +1,5 @@ from sentry.mediators.external_issues.creator import Creator -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.sentry_apps.services.app import app_service from sentry.testutils.cases import TestCase diff --git a/tests/sentry/mediators/external_issues/test_issue_link_creator.py b/tests/sentry/mediators/external_issues/test_issue_link_creator.py index edce90af8d8328..01e923118cd2cb 100644 --- a/tests/sentry/mediators/external_issues/test_issue_link_creator.py +++ b/tests/sentry/mediators/external_issues/test_issue_link_creator.py @@ -3,7 +3,7 @@ from sentry.coreapi import APIUnauthorized from sentry.mediators.external_issues.issue_link_creator import IssueLinkCreator -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.sentry_apps.services.app import app_service from sentry.testutils.cases import TestCase from sentry.users.services.user.serial import serialize_rpc_user diff --git a/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issue_actions.py b/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issue_actions.py index 2f7947175c37bc..79f7a75374dc52 100644 --- a/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issue_actions.py +++ b/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issue_actions.py @@ -1,7 +1,7 @@ import responses from django.urls import reverse -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.testutils.cases import APITestCase diff --git a/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issue_details.py b/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issue_details.py index fa606a8293d52a..581161b2900861 100644 --- a/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issue_details.py +++ b/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issue_details.py @@ -1,4 +1,4 @@ -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.testutils.cases import APITestCase diff --git a/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issues.py b/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issues.py index 964e2646569c06..7cc34b128028ef 100644 --- a/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issues.py +++ b/tests/sentry/sentry_apps/api/endpoints/test_sentry_app_installation_external_issues.py @@ -1,6 +1,6 @@ from django.urls import reverse -from sentry.models.platformexternalissue import PlatformExternalIssue +from sentry.sentry_apps.models.platformexternalissue import PlatformExternalIssue from sentry.testutils.cases import APITestCase
5ff970640c2e9d8ddb9b9003cd3b6cbc6059c37d
2024-10-25 23:21:00
anthony sottile
ref: fix typing for organization_teams (#79750)
false
fix typing for organization_teams (#79750)
ref
diff --git a/pyproject.toml b/pyproject.toml index b031fc4a9a1414..25f96f74e0600c 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -155,7 +155,6 @@ module = [ "sentry.api.endpoints.organization_search_details", "sentry.api.endpoints.organization_sessions", "sentry.api.endpoints.organization_stats", - "sentry.api.endpoints.organization_teams", "sentry.api.endpoints.project_index", "sentry.api.endpoints.project_ownership", "sentry.api.endpoints.project_release_files", diff --git a/src/sentry/api/endpoints/organization_teams.py b/src/sentry/api/endpoints/organization_teams.py index 5b7f1f6bea4562..0a43e883293e19 100644 --- a/src/sentry/api/endpoints/organization_teams.py +++ b/src/sentry/api/endpoints/organization_teams.py @@ -131,16 +131,18 @@ def get(self, request: Request, organization) -> Response: ) elif key == "query": - value = " ".join(value) - queryset = queryset.filter(Q(name__icontains=value) | Q(slug__icontains=value)) + joined_value = " ".join(value) + queryset = queryset.filter( + Q(name__icontains=joined_value) | Q(slug__icontains=joined_value) + ) elif key == "slug": queryset = queryset.filter(slug__in=value) elif key == "id": try: - value = [int(item) for item in value] + int_values = [int(item) for item in value] except ValueError: raise ParseError(detail="Invalid id value") - queryset = queryset.filter(id__in=value) + queryset = queryset.filter(id__in=int_values) else: queryset = queryset.none()
14b4c8279102e19304e2ccc80690a0937a4175c1
2024-08-30 01:29:00
Scott Cooper
ref(ui): Remove unreachable ?? paths (#76744)
false
Remove unreachable ?? paths (#76744)
ref
diff --git a/static/app/components/forms/model.tsx b/static/app/components/forms/model.tsx index 6711f0373ed543..230d68a47ab7d5 100644 --- a/static/app/components/forms/model.tsx +++ b/static/app/components/forms/model.tsx @@ -241,7 +241,7 @@ class FormModel { * Set form options */ setFormOptions(options: FormOptions) { - this.options = {...this.options, ...options} || {}; + this.options = {...this.options, ...options}; } /** diff --git a/static/app/views/insights/browser/common/queries/useResourcesQuery.ts b/static/app/views/insights/browser/common/queries/useResourcesQuery.ts index 82deeb275623e6..7ce7f2f389a408 100644 --- a/static/app/views/insights/browser/common/queries/useResourcesQuery.ts +++ b/static/app/views/insights/browser/common/queries/useResourcesQuery.ts @@ -169,9 +169,7 @@ export const getResourceTypeFilter = ( let resourceFilter: string[] = [`${SPAN_OP}:resource.*`]; if (selectedSpanOp) { - resourceFilter = [SPAN_OP_FILTER[selectedSpanOp].join(' OR ')] || [ - `${SPAN_OP}:${selectedSpanOp}`, - ]; + resourceFilter = [SPAN_OP_FILTER[selectedSpanOp].join(' OR ')]; } else if (defaultResourceTypes) { resourceFilter = [ defaultResourceTypes.map(type => SPAN_OP_FILTER[type].join(' OR ')).join(' OR '), diff --git a/static/app/views/insights/browser/webVitals/queries/storedScoreQueries/useProjectWebVitalsScoresTimeseriesQuery.tsx b/static/app/views/insights/browser/webVitals/queries/storedScoreQueries/useProjectWebVitalsScoresTimeseriesQuery.tsx index 28b888a769d898..346c69c2ce0583 100644 --- a/static/app/views/insights/browser/webVitals/queries/storedScoreQueries/useProjectWebVitalsScoresTimeseriesQuery.tsx +++ b/static/app/views/insights/browser/webVitals/queries/storedScoreQueries/useProjectWebVitalsScoresTimeseriesQuery.tsx @@ -136,7 +136,7 @@ export const useProjectWebVitalsScoresTimeseriesQuery = ({ data[webVital].push({ value: result?.data?.[`weighted_performance_score(measurements.score.${webVital})`] - ?.data[index][1][0].count * 100 ?? 0, + ?.data[index][1][0].count * 100, name: interval[0] * 1000, }); }); @@ -148,7 +148,7 @@ export const useProjectWebVitalsScoresTimeseriesQuery = ({ value: result?.data?.[`performance_score(measurements.score.${webVital})`]?.data[ index - ][1][0].count * 100 ?? 0, + ][1][0].count * 100, name: interval[0] * 1000, }); }); diff --git a/static/app/views/insights/queues/charts/latencyChart.tsx b/static/app/views/insights/queues/charts/latencyChart.tsx index 0d20241b9e3918..84fe0b4ec767c9 100644 --- a/static/app/views/insights/queues/charts/latencyChart.tsx +++ b/static/app/views/insights/queues/charts/latencyChart.tsx @@ -28,18 +28,16 @@ export function LatencyChart({error, destination, referrer}: Props) { top: '8px', bottom: '0', }} - data={ - [ - { - seriesName: t('Average Time in Queue'), - data: data['avg(messaging.message.receive.latency)'].data, - }, - { - seriesName: t('Average Processing Time'), - data: data['avg(span.duration)'].data, - }, - ] ?? [] - } + data={[ + { + seriesName: t('Average Time in Queue'), + data: data['avg(messaging.message.receive.latency)'].data, + }, + { + seriesName: t('Average Processing Time'), + data: data['avg(span.duration)'].data, + }, + ]} loading={isPending} error={error} chartColors={CHART_PALETTE[2].slice(1)} diff --git a/static/app/views/insights/queues/charts/throughputChart.tsx b/static/app/views/insights/queues/charts/throughputChart.tsx index 6604069b493a4e..857489ccd946ca 100644 --- a/static/app/views/insights/queues/charts/throughputChart.tsx +++ b/static/app/views/insights/queues/charts/throughputChart.tsx @@ -36,18 +36,16 @@ export function ThroughputChart({error, destination, referrer}: Props) { top: '8px', bottom: '0', }} - data={ - [ - { - seriesName: t('Published'), - data: publishData['spm()'].data, - }, - { - seriesName: t('Processed'), - data: processData['spm()'].data, - }, - ] ?? [] - } + data={[ + { + seriesName: t('Published'), + data: publishData['spm()'].data, + }, + { + seriesName: t('Processed'), + data: processData['spm()'].data, + }, + ]} loading={isPublishDataLoading || isProcessDataLoading} error={error} chartColors={CHART_PALETTE[2].slice(1, 3)} diff --git a/static/app/views/performance/newTraceDetails/traceRenderers/traceTextMeasurer.tsx b/static/app/views/performance/newTraceDetails/traceRenderers/traceTextMeasurer.tsx index ab6ea4cdc6ecf6..b5bdcbbf8cfc1d 100644 --- a/static/app/views/performance/newTraceDetails/traceRenderers/traceTextMeasurer.tsx +++ b/static/app/views/performance/newTraceDetails/traceRenderers/traceTextMeasurer.tsx @@ -28,8 +28,8 @@ export class TraceTextMeasurer { return; } - canvas.width = 50 * window.devicePixelRatio ?? 1; - canvas.height = 50 * window.devicePixelRatio ?? 1; + canvas.width = 50 * window.devicePixelRatio; + canvas.height = 50 * window.devicePixelRatio; ctx.font = '11px' + theme.text.family; diff --git a/static/app/views/projectDetail/charts/projectSessionsChartRequest.tsx b/static/app/views/projectDetail/charts/projectSessionsChartRequest.tsx index 4182933a9c3725..ce30e2b87add2d 100644 --- a/static/app/views/projectDetail/charts/projectSessionsChartRequest.tsx +++ b/static/app/views/projectDetail/charts/projectSessionsChartRequest.tsx @@ -256,7 +256,7 @@ class ProjectSessionsChartRequest extends Component< const crashedSessionsPercent = responseData.groups[0]?.series[field].slice( fetchedWithPrevious ? dataMiddleIndex : 0 - )[i] * 100 ?? 0; + )[i] * 100; return { name: interval, @@ -271,8 +271,7 @@ class ProjectSessionsChartRequest extends Component< seriesName: t('Previous Period'), data: responseData.intervals.slice(0, dataMiddleIndex).map((_interval, i) => { const crashedSessionsPercent = - responseData.groups[0]?.series[field].slice(0, dataMiddleIndex)[i] * 100 ?? - 0; + responseData.groups[0]?.series[field].slice(0, dataMiddleIndex)[i] * 100; return { name: responseData.intervals[i + dataMiddleIndex], diff --git a/webpack.config.ts b/webpack.config.ts index 3dbcf93a3324ad..b01385ff20528a 100644 --- a/webpack.config.ts +++ b/webpack.config.ts @@ -98,7 +98,7 @@ const SENTRY_EXPERIMENTAL_SPA = const SENTRY_SPA_DSN = SENTRY_EXPERIMENTAL_SPA ? env.SENTRY_SPA_DSN : undefined; const CODECOV_TOKEN = env.CODECOV_TOKEN; // value should come back as either 'true' or 'false' or undefined -const ENABLE_CODECOV_BA = env.CODECOV_ENABLE_BA === 'true' ?? false; +const ENABLE_CODECOV_BA = env.CODECOV_ENABLE_BA === 'true'; // this is the path to the django "sentry" app, we output the webpack build here to `dist` // so that `django collectstatic` and so that we can serve the post-webpack bundles
b4a36d63b93e620073f3ab7b52284116c5232863
2023-03-07 00:13:17
Snigdha Sharma
feat(codecov-v2): Add flag to restrict by plan (#45407)
false
Add flag to restrict by plan (#45407)
feat
diff --git a/src/sentry/conf/server.py b/src/sentry/conf/server.py index 914e78ac542aa4..443a18f83b527e 100644 --- a/src/sentry/conf/server.py +++ b/src/sentry/conf/server.py @@ -1022,6 +1022,8 @@ def SOCIAL_AUTH_DEFAULT_USERNAME(): "organizations:codecov-stacktrace-integration-v2": False, # Enables the cron job to auto-enable codecov integrations. "organizations:auto-enable-codecov": False, + # The overall flag for codecov integration, gated by plans. + "organizations:codecov-integration": False, # Enables getting commit sha from git blame for codecov. "organizations:codecov-commit-sha-from-git-blame": False, # Enables automatically deriving of code mappings diff --git a/src/sentry/features/__init__.py b/src/sentry/features/__init__.py index acbdd208667210..b45544b27be262 100644 --- a/src/sentry/features/__init__.py +++ b/src/sentry/features/__init__.py @@ -227,6 +227,7 @@ default_manager.add("organizations:derive-code-mappings-dry-run", OrganizationFeature) default_manager.add("organizations:codecov-stacktrace-integration", OrganizationFeature, True) default_manager.add("organizations:codecov-stacktrace-integration-v2", OrganizationFeature, True) +default_manager.add("organizations:codecov-integration", OrganizationFeature) default_manager.add("organizations:auto-enable-codecov", OrganizationFeature) default_manager.add("organizations:codecov-commit-sha-from-git-blame", OrganizationFeature, True) default_manager.add("organizations:artifact-bundles", OrganizationFeature, True)
712d612833eb48b3773d4dc22a34883d432ac2a8
2019-07-10 21:12:46
Alberto Leal
fix: Pass the correct query to the /organizations/:org-slug/events-meta/ endpoint (#13917)
false
Pass the correct query to the /organizations/:org-slug/events-meta/ endpoint (#13917)
fix
diff --git a/src/sentry/static/sentry/app/__mocks__/api.jsx b/src/sentry/static/sentry/app/__mocks__/api.jsx index 3f47a12852582c..67f0258bcfd9e3 100644 --- a/src/sentry/static/sentry/app/__mocks__/api.jsx +++ b/src/sentry/static/sentry/app/__mocks__/api.jsx @@ -12,6 +12,10 @@ const respond = (isAsync, fn, ...args) => { } }; +const DEFAULT_MOCK_RESPONSE_OPTIONS = { + predicate: () => true, +}; + class Client { static mockResponses = []; @@ -20,7 +24,7 @@ class Client { } // Returns a jest mock that represents Client.request calls - static addMockResponse(response) { + static addMockResponse(response, options = DEFAULT_MOCK_RESPONSE_OPTIONS) { const mock = jest.fn(); Client.mockResponses.unshift([ { @@ -32,14 +36,19 @@ class Client { headers: response.headers || {}, }, mock, + options.predicate, ]); return mock; } static findMockResponse(url, options) { - return Client.mockResponses.find(([response]) => { - return url === response.url && (options.method || 'GET') === response.method; + return Client.mockResponses.find(([response, mock, predicate]) => { + const matchesURL = url === response.url; + const matchesMethod = (options.method || 'GET') === response.method; + const matchesPredicate = predicate(url, options); + + return matchesURL && matchesMethod && matchesPredicate; }); } diff --git a/src/sentry/static/sentry/app/views/organizationEventsV2/tags.jsx b/src/sentry/static/sentry/app/views/organizationEventsV2/tags.jsx index 196c18fe31d690..680a378c7e61a6 100644 --- a/src/sentry/static/sentry/app/views/organizationEventsV2/tags.jsx +++ b/src/sentry/static/sentry/app/views/organizationEventsV2/tags.jsx @@ -8,7 +8,12 @@ import SentryTypes from 'app/sentryTypes'; import Placeholder from 'app/components/placeholder'; import TagDistributionMeter from 'app/components/tagDistributionMeter'; import withApi from 'app/utils/withApi'; -import {fetchTagDistribution, fetchTotalCount, getEventTagSearchUrl} from './utils'; +import { + fetchTagDistribution, + fetchTotalCount, + getEventTagSearchUrl, + getQuery, +} from './utils'; import {MODAL_QUERY_KEYS} from './data'; class Tags extends React.Component { @@ -61,7 +66,11 @@ class Tags extends React.Component { }); try { - const totalValues = await fetchTotalCount(api, organization.slug, location.query); + const totalValues = await fetchTotalCount( + api, + organization.slug, + getQuery(view, location) + ); this.setState({totalValues}); } catch (err) { Sentry.captureException(err); diff --git a/src/sentry/static/sentry/app/views/organizationEventsV2/utils.jsx b/src/sentry/static/sentry/app/views/organizationEventsV2/utils.jsx index 20d7bb08bdb1de..5e97e79f94f231 100644 --- a/src/sentry/static/sentry/app/views/organizationEventsV2/utils.jsx +++ b/src/sentry/static/sentry/app/views/organizationEventsV2/utils.jsx @@ -1,4 +1,4 @@ -import {pick} from 'lodash'; +import {pick, get} from 'lodash'; import {DEFAULT_PER_PAGE} from 'app/constants'; import {URL_PARAM} from 'app/constants/globalSelectionHeader'; @@ -25,7 +25,9 @@ export function getQuery(view, location) { const fields = []; const groupby = view.data.groupby ? [...view.data.groupby] : []; - view.data.fields.forEach(field => { + const viewFields = get(view, 'data.fields', []); + + viewFields.forEach(field => { if (SPECIAL_FIELDS.hasOwnProperty(field)) { const specialField = SPECIAL_FIELDS[field]; diff --git a/tests/js/spec/views/organizationEventsV2/tags.spec.jsx b/tests/js/spec/views/organizationEventsV2/tags.spec.jsx index 6fe914b7367c9b..43dfad776b78da 100644 --- a/tests/js/spec/views/organizationEventsV2/tags.spec.jsx +++ b/tests/js/spec/views/organizationEventsV2/tags.spec.jsx @@ -7,15 +7,39 @@ import {Tags} from 'app/views/organizationEventsV2/tags'; describe('Tags', function() { const org = TestStubs.Organization(); beforeEach(function() { - Client.addMockResponse({ - url: `/organizations/${org.slug}/events-distribution/`, - body: { - key: 'release', - name: 'Release', - totalValues: 2, - topValues: [{count: 2, value: 'abcd123', name: 'abcd123'}], + Client.addMockResponse( + { + url: `/organizations/${org.slug}/events-distribution/`, + body: { + key: 'release', + name: 'Release', + totalValues: 2, + topValues: [{count: 2, value: 'abcd123', name: 'abcd123'}], + }, }, - }); + { + predicate: (url, options) => { + return options.query.key === 'release'; + }, + } + ); + + Client.addMockResponse( + { + url: `/organizations/${org.slug}/events-distribution/`, + body: { + key: 'environment', + name: 'Environment', + totalValues: 2, + topValues: [{count: 2, value: 'abcd123', name: 'abcd123'}], + }, + }, + { + predicate: (url, options) => { + return options.query.key === 'environment'; + }, + } + ); Client.addMockResponse({ url: `/organizations/${org.slug}/events-meta/`, @@ -47,9 +71,13 @@ describe('Tags', function() { /> ); + // component is in loading state expect(wrapper.find('StyledPlaceholder')).toHaveLength(2); + await tick(); wrapper.update(); + + // component has loaded expect(wrapper.find('StyledPlaceholder')).toHaveLength(0); }); });
72df0cab9b215b281ad1f5110403108237754a2b
2021-01-14 06:03:06
Danny Lee
ref(ui): Pixel matters on SettingsPageHeader (#23074)
false
Pixel matters on SettingsPageHeader (#23074)
ref
diff --git a/src/sentry/static/sentry/app/views/settings/components/settingsPageHeader.tsx b/src/sentry/static/sentry/app/views/settings/components/settingsPageHeader.tsx index 3c7d60af1e769b..da304bc8f233e1 100644 --- a/src/sentry/static/sentry/app/views/settings/components/settingsPageHeader.tsx +++ b/src/sentry/static/sentry/app/views/settings/components/settingsPageHeader.tsx @@ -61,9 +61,8 @@ class UnstyledSettingsPageHeader extends React.Component<Props> { {action && <Action isNarrow={isNarrow}>{action}</Action>} </TitleAndActions> - {body} - - {tabs} + {body && <BodyWrapper>{body}</BodyWrapper>} + {tabs && <TabsWrapper>{tabs}</TabsWrapper>} </div> ); } @@ -84,13 +83,13 @@ const TitleWrapper = styled('div')` const Title = styled('div')<TitleProps & React.HTMLProps<HTMLDivElement>>` ${p => !p.styled && `font-size: 20px; font-weight: 600;`}; - margin: ${space(4)} ${space(2)} ${space(4)} 0; + margin: ${space(4)} ${space(2)} ${space(3)} 0; `; const Subtitle = styled('div')` color: ${p => p.theme.gray400}; font-weight: 400; font-size: ${p => p.theme.fontSizeLarge}; - padding: ${space(1.5)} 0 ${space(3)}; + padding: ${space(1.5)} 0 0; `; const Icon = styled('div')` @@ -108,4 +107,13 @@ const SettingsPageHeader = styled(UnstyledSettingsPageHeader)< margin-top: -${space(4)}; `; +const BodyWrapper = styled('div')` + flex: 1; + margin: 0 0 ${space(3)}; +`; +const TabsWrapper = styled('div')` + flex: 1; + margin: 0; /* sentry/components/navTabs has added margin */ +`; + export default SettingsPageHeader;
db043e06bdd6111e56fc8ee0552d8bce2efe9053
2025-02-15 03:10:54
George Gritsouk
ref(dashboards): Namespaced `Widget` components (#85238)
false
Namespaced `Widget` components (#85238)
ref
diff --git a/static/app/views/dashboards/widgets/bigNumberWidget/bigNumberWidget.tsx b/static/app/views/dashboards/widgets/bigNumberWidget/bigNumberWidget.tsx index e3cc6f83e8d328..c769fd82ff17e0 100644 --- a/static/app/views/dashboards/widgets/bigNumberWidget/bigNumberWidget.tsx +++ b/static/app/views/dashboards/widgets/bigNumberWidget/bigNumberWidget.tsx @@ -13,8 +13,6 @@ import { import {DEFAULT_FIELD, MISSING_DATA_MESSAGE} from '../common/settings'; import type {StateProps} from '../common/types'; -import {DEEMPHASIS_COLOR_NAME, LOADING_PLACEHOLDER} from './settings'; - export interface BigNumberWidgetProps extends StateProps, Omit<WidgetFrameProps, 'children'>, @@ -32,7 +30,7 @@ export function BigNumberWidget(props: BigNumberWidgetProps) { revealActions={props.revealActions} revealTooltip={props.revealTooltip} > - <LoadingPlaceholder>{LOADING_PLACEHOLDER}</LoadingPlaceholder> + <BigNumberWidgetVisualization.LoadingPlaceholder /> </WidgetFrame> ); } @@ -82,8 +80,3 @@ const BigNumberResizeWrapper = styled('div')` position: relative; flex-grow: 1; `; - -const LoadingPlaceholder = styled('span')` - color: ${p => p.theme[DEEMPHASIS_COLOR_NAME]}; - font-size: ${p => p.theme.fontSizeLarge}; -`; diff --git a/static/app/views/dashboards/widgets/bigNumberWidget/bigNumberWidgetVisualization.tsx b/static/app/views/dashboards/widgets/bigNumberWidget/bigNumberWidgetVisualization.tsx index b3d69f2bf7380a..1a2fb48c961d82 100644 --- a/static/app/views/dashboards/widgets/bigNumberWidget/bigNumberWidgetVisualization.tsx +++ b/static/app/views/dashboards/widgets/bigNumberWidget/bigNumberWidgetVisualization.tsx @@ -17,6 +17,7 @@ import type { import {NON_FINITE_NUMBER_MESSAGE} from '../common/settings'; +import {DEEMPHASIS_COLOR_NAME, LOADING_PLACEHOLDER} from './settings'; import {ThresholdsIndicator} from './thresholdsIndicator'; export interface BigNumberWidgetVisualizationProps { @@ -175,3 +176,12 @@ const NumberContainerOverride = styled('div')` white-space: nowrap; } `; + +const LoadingPlaceholder = styled('span')` + color: ${p => p.theme[DEEMPHASIS_COLOR_NAME]}; + font-size: ${p => p.theme.fontSizeLarge}; +`; + +BigNumberWidgetVisualization.LoadingPlaceholder = function () { + return <LoadingPlaceholder>{LOADING_PLACEHOLDER}</LoadingPlaceholder>; +}; diff --git a/static/app/views/dashboards/widgets/common/widgetFrame.tsx b/static/app/views/dashboards/widgets/common/widgetFrame.tsx index 57af0ac35c4d30..8c32b98e692db5 100644 --- a/static/app/views/dashboards/widgets/common/widgetFrame.tsx +++ b/static/app/views/dashboards/widgets/common/widgetFrame.tsx @@ -1,21 +1,16 @@ import {Fragment} from 'react'; +import styled from '@emotion/styled'; import type {BadgeProps} from 'sentry/components/badge/badge'; -import {LinkButton} from 'sentry/components/button'; +import Badge from 'sentry/components/badge/badge'; +import {Button, LinkButton} from 'sentry/components/button'; import {DropdownMenu, type MenuItemProps} from 'sentry/components/dropdownMenu'; import {Tooltip} from 'sentry/components/tooltip'; import {IconEllipsis, IconExpand, IconWarning} from 'sentry/icons'; import {t} from 'sentry/locale'; -import {ErrorPanel} from '../widgetLayout/errorPanel'; -import {WidgetBadge} from '../widgetLayout/widgetBadge'; -import {WidgetButton} from '../widgetLayout/widgetButton'; -import { - WidgetDescription, - type WidgetDescriptionProps, -} from '../widgetLayout/widgetDescription'; -import {WidgetLayout} from '../widgetLayout/widgetLayout'; -import {WidgetTitle} from '../widgetLayout/widgetTitle'; +import {Widget} from '../widget/widget'; +import type {WidgetDescriptionProps} from '../widget/widgetDescription'; import {TooltipIconTrigger} from './tooltipIconTrigger'; import type {StateProps} from './types'; @@ -59,7 +54,7 @@ export function WidgetFrame(props: WidgetFrameProps) { const shouldShowActions = actions && actions.length > 0; return ( - <WidgetLayout + <Widget ariaLabel="Widget panel" borderless={props.borderless} Title={ @@ -72,7 +67,7 @@ export function WidgetFrame(props: WidgetFrameProps) { </Tooltip> )} - <WidgetTitle title={props.title} /> + <Widget.WidgetTitle title={props.title} /> {props.badgeProps && (Array.isArray(props.badgeProps) ? props.badgeProps : [props.badgeProps]).map( @@ -87,7 +82,7 @@ export function WidgetFrame(props: WidgetFrameProps) { <Fragment> {props.description && ( // Ideally we'd use `QuestionTooltip` but we need to firstly paint the icon dark, give it 100% opacity, and remove hover behaviour. - <WidgetDescription + <Widget.WidgetDescription title={props.title} description={props.description} revealTooltip={props.revealTooltip ?? 'hover'} @@ -110,12 +105,13 @@ export function WidgetFrame(props: WidgetFrameProps) { {actions[0]!.label} </LinkButton> ) : ( - <WidgetButton + <Button + size="xs" disabled={props.actionsDisabled} onClick={actions[0]!.onAction} > {actions[0]!.label} - </WidgetButton> + </Button> ) ) : null} @@ -137,7 +133,8 @@ export function WidgetFrame(props: WidgetFrameProps) { )} {shouldShowFullScreenViewButton && ( - <WidgetButton + <Button + size="xs" aria-label={t('Open Full-Screen View')} borderless icon={<IconExpand />} @@ -148,7 +145,7 @@ export function WidgetFrame(props: WidgetFrameProps) { )} </Fragment> } - Visualization={props.error ? <ErrorPanel error={error} /> : props.children} + Visualization={props.error ? <Widget.WidgetError error={error} /> : props.children} noVisualizationPadding={props.noVisualizationPadding} /> ); @@ -160,6 +157,10 @@ interface TitleActionsProps { disabledMessage: string; } +const WidgetBadge = styled(Badge)` + flex-shrink: 0; +`; + function TitleActionsWrapper({disabled, disabledMessage, children}: TitleActionsProps) { if (!disabled || !disabledMessage) { return children; diff --git a/static/app/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidget.tsx b/static/app/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidget.tsx index 75bacf5530c3fe..8e94707fd138e5 100644 --- a/static/app/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidget.tsx +++ b/static/app/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidget.tsx @@ -10,7 +10,6 @@ import { import {MISSING_DATA_MESSAGE} from '../common/settings'; import type {StateProps} from '../common/types'; -import {LoadingPanel} from '../widgetLayout/loadingPanel'; export interface TimeSeriesWidgetProps extends StateProps, @@ -30,7 +29,7 @@ export function TimeSeriesWidget(props: TimeSeriesWidgetProps) { revealActions={props.revealActions} revealTooltip={props.revealTooltip} > - <LoadingPanel /> + <TimeSeriesWidgetVisualization.LoadingPlaceholder /> </WidgetFrame> ); } diff --git a/static/app/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidgetVisualization.tsx b/static/app/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidgetVisualization.tsx index 506da1c5cbd00a..e8a0db6f028524 100644 --- a/static/app/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidgetVisualization.tsx +++ b/static/app/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidgetVisualization.tsx @@ -1,6 +1,7 @@ import {useRef} from 'react'; import {useNavigate} from 'react-router-dom'; import {useTheme} from '@emotion/react'; +import styled from '@emotion/styled'; import type {BarSeriesOption, LineSeriesOption} from 'echarts'; import type { TooltipFormatterCallback, @@ -11,8 +12,10 @@ import type EChartsReactCore from 'echarts-for-react/lib/core'; import BaseChart from 'sentry/components/charts/baseChart'; import {getFormatter} from 'sentry/components/charts/components/tooltip'; import LineSeries from 'sentry/components/charts/series/lineSeries'; +import TransparentLoadingMask from 'sentry/components/charts/transparentLoadingMask'; import {useChartZoom} from 'sentry/components/charts/useChartZoom'; import {isChartHovered, truncationFormatter} from 'sentry/components/charts/utils'; +import LoadingIndicator from 'sentry/components/loadingIndicator'; import type {Series} from 'sentry/types/echarts'; import {defined} from 'sentry/utils'; import {uniq} from 'sentry/utils/array/uniq'; @@ -27,7 +30,7 @@ import useOrganization from 'sentry/utils/useOrganization'; import usePageFilters from 'sentry/utils/usePageFilters'; import {useWidgetSyncContext} from '../../contexts/widgetSyncContext'; -import {NO_PLOTTABLE_VALUES} from '../common/settings'; +import {NO_PLOTTABLE_VALUES, X_GUTTER, Y_GUTTER} from '../common/settings'; import type {Aliases, Release, TimeSeries, TimeseriesSelection} from '../common/types'; import {BarChartWidgetSeries} from './seriesConstructors/barChartWidgetSeries'; @@ -351,4 +354,30 @@ export function TimeSeriesWidgetVisualization(props: TimeSeriesWidgetVisualizati ); } +function LoadingPanel() { + return ( + <LoadingPlaceholder> + <LoadingMask visible /> + <LoadingIndicator mini /> + </LoadingPlaceholder> + ); +} + +const LoadingPlaceholder = styled('div')` + position: absolute; + inset: 0; + + display: flex; + justify-content: center; + align-items: center; + + padding: ${Y_GUTTER} ${X_GUTTER}; +`; + +const LoadingMask = styled(TransparentLoadingMask)` + background: ${p => p.theme.background}; +`; + +TimeSeriesWidgetVisualization.LoadingPlaceholder = LoadingPanel; + const GLOBAL_STACK_NAME = 'time-series-visualization-widget-stack'; diff --git a/static/app/views/dashboards/widgets/widget/widget.stories.tsx b/static/app/views/dashboards/widgets/widget/widget.stories.tsx new file mode 100644 index 00000000000000..209daa65b4d41b --- /dev/null +++ b/static/app/views/dashboards/widgets/widget/widget.stories.tsx @@ -0,0 +1,152 @@ +import {Fragment} from 'react'; +import styled from '@emotion/styled'; + +import {Button} from 'sentry/components/button'; +import {CodeSnippet} from 'sentry/components/codeSnippet'; +import JSXNode from 'sentry/components/stories/jsxNode'; +import SizingWindow from 'sentry/components/stories/sizingWindow'; +import storyBook from 'sentry/stories/storyBook'; + +import {sampleDurationTimeSeries} from '../lineChartWidget/fixtures/sampleDurationTimeSeries'; +import {TimeSeriesWidgetVisualization} from '../timeSeriesWidget/timeSeriesWidgetVisualization'; + +import {Widget} from './widget'; + +// eslint-disable-next-line import/no-webpack-loader-syntax +import types from '!!type-loader!sentry/views/dashboards/widgets/widget/widget'; + +export default storyBook('Widget', (story, APIReference) => { + APIReference(types.exported); + + story('Getting Started', () => { + const isLoading: boolean = false; + const hasError: boolean = false; + + return ( + <Fragment> + <p> + A "Widget" is a common UI element in Sentry, used to show aggregate data. It + consists of a frame, a title, an optional description, a visualization (e.g., a + line chart), and an optional footer. + </p> + + <SmallSizingWindow> + <Widget + Title={<Widget.WidgetTitle title="epm() : /insights/frontend/assets" />} + Actions={ + <Widget.WidgetToolbar> + <Button size="xs">Say More</Button> + <Button size="xs">Say Less</Button> + <Widget.WidgetDescription + title="epm() : /insights/frontend/assets" + description="Events received, tracked per minute" + /> + </Widget.WidgetToolbar> + } + Visualization={ + isLoading ? ( + <TimeSeriesWidgetVisualization.LoadingPlaceholder /> + ) : hasError ? ( + <Widget.WidgetError error="Oh no!" /> + ) : ( + <TimeSeriesWidgetVisualization + visualizationType="line" + timeSeries={[sampleDurationTimeSeries]} + /> + ) + } + Footer={<span>This data is incomplete!</span>} + /> + </SmallSizingWindow> + </Fragment> + ); + }); + + story('Widget', () => { + return ( + <Fragment> + <p> + <JSXNode name="Widget" /> is a component as well as a component namespace. It's + designed to make it easy to create widgets of your own.{' '} + </p> + <p> + <JSXNode name="Widget" /> is a layout-only component. It contains no logic, all + it does it place the passed sub-components in correct locations in a bordered + widget frame. The contents of the <code>Title</code> prop are shown in the top + left, and are always visible. The title is truncated to fit. The contents of the{' '} + <code>Actions</code> prop are shown in the top right, and only shown on hover. + You can set the <code>revealActions</code> prop to <code>"always"</code> to + always show the actions. Actions are not truncated. The contents of{' '} + <code>Visualization</code> are always visible, shown below the title and + actions. The layout expands both horizontally and vertically to fit the parent. + </p> + + <p> + <JSXNode name="Widget" /> also provides a few sub-components: + <ul> + <li> + <JSXNode name="Widget.TextTitle" /> is a truncated title string + </li> + <li> + <JSXNode name="Widget.Description" /> is a description tooltip + </li> + <li> + <JSXNode name="Widget.Toolbar" /> is a wrapper for multiple buttons + </li> + <li> + <JSXNode name="Widget.Error" /> is an error panel that takes over the{' '} + <code>Visualization</code> if needed + </li> + </ul> + </p> + + <p> + The best way to illustrate these concepts is a full example that uses all of the + available components. The code below describes how to render the example widget + shown above. + </p> + + <CodeSnippet language="jsx"> + {`import {LineChartWidgetVisualization} from '../lineChartWidget/lineChartWidgetVisualization'; +import {sampleDurationTimeSeries} from '../lineChartWidget/fixtures/sampleDurationTimeSeries'; + +import {Widget} from './widget'; + +<Widget + Title={<Widget.TextTitle title="epm() : /insights/frontend/assets" />} + Actions={ + <Widget.Toolbar> + <Button size="xs">Say More</Button> + <Button size="xs">Say Less</Button> + <Widget.Description + title="epm() : /insights/frontend/assets" + description="Events received, tracked per minute" + /> + </Widget.Toolbar> + } + Visualization={ + isLoading ? ( + <TimeSeriesWidgetVisualization.Placeholder /> + ) : hasError ? ( + <Widget.Error error="Oh no!" /> + ) : ( + <TimeSeriesWidgetVisualization + visualizationType="line" + timeSeries={[sampleDurationTimeSeries]} + /> + ) + } + Footer={<span>This data is incomplete!</span>} +/> + + `} + </CodeSnippet> + </Fragment> + ); + }); +}); + +const SmallSizingWindow = styled(SizingWindow)` + width: 400px; + height: 300px; +`; diff --git a/static/app/views/dashboards/widgets/widgetLayout/widgetLayout.tsx b/static/app/views/dashboards/widgets/widget/widget.tsx similarity index 70% rename from static/app/views/dashboards/widgets/widgetLayout/widgetLayout.tsx rename to static/app/views/dashboards/widgets/widget/widget.tsx index 5b3228e8f07856..118d85bcdb701f 100644 --- a/static/app/views/dashboards/widgets/widgetLayout/widgetLayout.tsx +++ b/static/app/views/dashboards/widgets/widget/widget.tsx @@ -7,23 +7,56 @@ import {space} from 'sentry/styles/space'; import {MIN_HEIGHT, MIN_WIDTH, X_GUTTER, Y_GUTTER} from '../common/settings'; -import {ErrorPanel} from './errorPanel'; - -export interface WidgetLayoutProps { +import {WidgetDescription} from './widgetDescription'; +import {WidgetError} from './widgetError'; +import {WidgetTitle} from './widgetTitle'; +import {WidgetToolbar} from './widgetToolbar'; + +export interface Widget { + /** + * Placed in the top right of the frame + */ Actions?: React.ReactNode; + /** + * Placed below the visualization, inside the frame + */ Footer?: React.ReactNode; + /** + * Placed in the top left of the frame + */ Title?: React.ReactNode; + /** + * Placed in the main area of the frame + */ Visualization?: React.ReactNode; ariaLabel?: string; + /** + * Removes frame border + */ borderless?: boolean; + /** + * Height in pixels. If omitted, the widget grows to fill the parent element. Avoid this! Setting the height via the parent element is more robust + */ height?: number; + /** + * Removes padding from the footer area + */ noFooterPadding?: boolean; + /** + * Removes padding from the header area + */ noHeaderPadding?: boolean; + /** + * Removes padding from the visualization area + */ noVisualizationPadding?: boolean; + /** + * If set to `"hover"`, the contents of the `Actions` slot are only shown on mouseover. If set to `"always"`, the contents of `Actions` are always shown + */ revealActions?: 'hover' | 'always'; } -export function WidgetLayout(props: WidgetLayoutProps) { +function WidgetLayout(props: Widget) { const {revealActions = 'hover'} = props; return ( @@ -41,7 +74,7 @@ export function WidgetLayout(props: WidgetLayoutProps) { {props.Visualization && ( <ErrorBoundary customComponent={({error}) => ( - <ErrorPanel error={error?.message ?? undefined} /> + <WidgetError error={error?.message ?? undefined} /> )} > <VisualizationWrapper noPadding={props.noVisualizationPadding}> @@ -57,6 +90,18 @@ export function WidgetLayout(props: WidgetLayoutProps) { ); } +// `Object.assign` ensures correct types by intersection the component with the +// extra properties. This allows rendering `<Widget>` as well as +// `<Widget.Description` and others +const exported = Object.assign(WidgetLayout, { + WidgetDescription, + WidgetTitle, + WidgetToolbar, + WidgetError, +}); + +export {exported as Widget}; + const HEADER_HEIGHT = '26px'; const TitleHoverItems = styled('div')` diff --git a/static/app/views/dashboards/widgets/widgetLayout/widgetDescription.tsx b/static/app/views/dashboards/widgets/widget/widgetDescription.tsx similarity index 73% rename from static/app/views/dashboards/widgets/widgetLayout/widgetDescription.tsx rename to static/app/views/dashboards/widgets/widget/widgetDescription.tsx index 4f94e1cf3644e6..cba1391db30020 100644 --- a/static/app/views/dashboards/widgets/widgetLayout/widgetDescription.tsx +++ b/static/app/views/dashboards/widgets/widget/widgetDescription.tsx @@ -16,18 +16,18 @@ export function WidgetDescription(props: WidgetDescriptionProps) { return ( <Tooltip title={ - <WidgetTooltipContents> - {props.title && <WidgetTooltipTitle>{props.title}</WidgetTooltipTitle>} + <TooltipContents> + {props.title && <TooltipTitle>{props.title}</TooltipTitle>} {props.description && ( - <WidgetTooltipDescription>{props.description}</WidgetTooltipDescription> + <TooltipDescription>{props.description}</TooltipDescription> )} - </WidgetTooltipContents> + </TooltipContents> } containerDisplayMode="grid" isHoverable forceVisible={props.revealTooltip === 'always' ? true : undefined} > - <WidgetTooltipButton + <TooltipButton aria-label={t('Widget description')} borderless size="xs" @@ -37,7 +37,7 @@ export function WidgetDescription(props: WidgetDescriptionProps) { ); } -const WidgetTooltipContents = styled('div')` +const TooltipContents = styled('div')` display: flex; flex-direction: column; gap: ${space(0.5)}; @@ -45,19 +45,19 @@ const WidgetTooltipContents = styled('div')` overflow: hidden; `; -const WidgetTooltipTitle = styled('div')` +const TooltipTitle = styled('div')` font-weight: bold; font-size: ${p => p.theme.fontSizeMedium}; text-align: left; `; -const WidgetTooltipDescription = styled('div')` +const TooltipDescription = styled('div')` font-size: ${p => p.theme.fontSizeSmall}; text-align: left; `; // We're using a button here to preserve tab accessibility -const WidgetTooltipButton = styled(Button)` +const TooltipButton = styled(Button)` pointer-events: none; padding-top: 0; padding-bottom: 0; diff --git a/static/app/views/dashboards/widgets/widgetLayout/errorPanel.tsx b/static/app/views/dashboards/widgets/widget/widgetError.tsx similarity index 92% rename from static/app/views/dashboards/widgets/widgetLayout/errorPanel.tsx rename to static/app/views/dashboards/widgets/widget/widgetError.tsx index 0d0dd9d46cfce5..e1290962bb7f23 100644 --- a/static/app/views/dashboards/widgets/widgetLayout/errorPanel.tsx +++ b/static/app/views/dashboards/widgets/widget/widgetError.tsx @@ -7,11 +7,11 @@ import type {StateProps} from 'sentry/views/dashboards/widgets/common/types'; import {X_GUTTER, Y_GUTTER} from '../common/settings'; -interface ErrorPanelProps { +interface WidgetErrorProps { error: StateProps['error']; } -export function ErrorPanel({error}: ErrorPanelProps) { +export function WidgetError({error}: WidgetErrorProps) { return ( <Panel> <NonShrinkingWarningIcon color={DEEMPHASIS_COLOR_NAME} size="md" /> diff --git a/static/app/views/dashboards/widgets/widgetLayout/widgetTitle.tsx b/static/app/views/dashboards/widgets/widget/widgetTitle.tsx similarity index 100% rename from static/app/views/dashboards/widgets/widgetLayout/widgetTitle.tsx rename to static/app/views/dashboards/widgets/widget/widgetTitle.tsx diff --git a/static/app/views/dashboards/widgets/widget/widgetToolbar.tsx b/static/app/views/dashboards/widgets/widget/widgetToolbar.tsx new file mode 100644 index 00000000000000..9e57d860b9953b --- /dev/null +++ b/static/app/views/dashboards/widgets/widget/widgetToolbar.tsx @@ -0,0 +1,8 @@ +import styled from '@emotion/styled'; + +import {space} from 'sentry/styles/space'; + +export const WidgetToolbar = styled('div')` + display: flex; + gap: ${space(0.5)}; +`; diff --git a/static/app/views/dashboards/widgets/widgetLayout/loadingPanel.tsx b/static/app/views/dashboards/widgets/widgetLayout/loadingPanel.tsx deleted file mode 100644 index e8b8bb83a9ea7b..00000000000000 --- a/static/app/views/dashboards/widgets/widgetLayout/loadingPanel.tsx +++ /dev/null @@ -1,29 +0,0 @@ -import styled from '@emotion/styled'; - -import TransparentLoadingMask from 'sentry/components/charts/transparentLoadingMask'; -import LoadingIndicator from 'sentry/components/loadingIndicator'; - -import {X_GUTTER, Y_GUTTER} from '../common/settings'; - -export function LoadingPanel() { - return ( - <LoadingPlaceholder> - <LoadingMask visible /> - <LoadingIndicator mini /> - </LoadingPlaceholder> - ); -} -const LoadingPlaceholder = styled('div')` - position: absolute; - inset: 0; - - display: flex; - justify-content: center; - align-items: center; - - padding: ${Y_GUTTER} ${X_GUTTER}; -`; - -const LoadingMask = styled(TransparentLoadingMask)` - background: ${p => p.theme.background}; -`; diff --git a/static/app/views/dashboards/widgets/widgetLayout/widgetBadge.tsx b/static/app/views/dashboards/widgets/widgetLayout/widgetBadge.tsx deleted file mode 100644 index f1ba67b8d511df..00000000000000 --- a/static/app/views/dashboards/widgets/widgetLayout/widgetBadge.tsx +++ /dev/null @@ -1,7 +0,0 @@ -import styled from '@emotion/styled'; - -import Badge from 'sentry/components/badge/badge'; - -export const WidgetBadge = styled(Badge)` - flex-shrink: 0; -`; diff --git a/static/app/views/dashboards/widgets/widgetLayout/widgetButton.tsx b/static/app/views/dashboards/widgets/widgetLayout/widgetButton.tsx deleted file mode 100644 index 3678286ad0313e..00000000000000 --- a/static/app/views/dashboards/widgets/widgetLayout/widgetButton.tsx +++ /dev/null @@ -1,11 +0,0 @@ -import type {ComponentProps} from 'react'; - -import {Button} from 'sentry/components/button'; - -export function WidgetButton(props: Omit<ComponentProps<typeof Button>, 'size'>) { - return ( - <Button {...props} size="xs"> - {props.children} - </Button> - ); -} diff --git a/static/app/views/dashboards/widgets/widgetLayout/widgetLayout.stories.tsx b/static/app/views/dashboards/widgets/widgetLayout/widgetLayout.stories.tsx deleted file mode 100644 index 8b7649e2e88259..00000000000000 --- a/static/app/views/dashboards/widgets/widgetLayout/widgetLayout.stories.tsx +++ /dev/null @@ -1,110 +0,0 @@ -import {Fragment} from 'react'; -import styled from '@emotion/styled'; - -import {CodeSnippet} from 'sentry/components/codeSnippet'; -import JSXNode from 'sentry/components/stories/jsxNode'; -import SizingWindow from 'sentry/components/stories/sizingWindow'; -import storyBook from 'sentry/stories/storyBook'; - -import {sampleDurationTimeSeries} from '../lineChartWidget/fixtures/sampleDurationTimeSeries'; -import {LineChartWidgetVisualization} from '../lineChartWidget/lineChartWidgetVisualization'; - -import {WidgetButton} from './widgetButton'; -import {WidgetDescription} from './widgetDescription'; -import {WidgetLayout} from './widgetLayout'; -import {WidgetTitle} from './widgetTitle'; - -export default storyBook('WidgetLayout', story => { - story('Getting Started', () => { - return ( - <Fragment> - <p> - In most cases, we recommend using standard widgets like{' '} - <JSXNode name="LineChartWidget" />. If this isn't possible (because of custom - layout needs), we offer a set of helper components. Components like{' '} - <JSXNode name="WidgetLayout" /> can be used to create a standard-looking widget - from near-scratch. - </p> - </Fragment> - ); - }); - - story('WidgetLayout', () => { - return ( - <Fragment> - <p> - <JSXNode name="WidgetLayout" /> is a layout-only component. It contains no - logic, all it does it place the passed components in correct locations in a - bordered widget frame. The contents of the <code>Title</code> prop are shown in - the top left, and are always visible. The title is truncated to fit. The - contents of the <code>Actions</code> prop are shown in the top right, and only - shown on hover. You can set the <code>revealActions</code> prop to{' '} - <code>"always"</code> to always show the actions. Actions are not truncated. The - contents of <code>Visualization</code> are always visible, shown below the title - and actions. The layout expands both horizontally and vertically to fit the - parent. - </p> - - <p> - In order to make a nice-looking custom widget layout we recommend using the - pre-built components that we provide alongside the layout. - </p> - - <CodeSnippet language="jsx"> - {`import {LineChartWidgetVisualization} from '../lineChartWidget/lineChartWidgetVisualization'; -import {sampleDurationTimeSeries} from '../lineChartWidget/fixtures/sampleDurationTimeSeries'; - -import {WidgetButton} from './widgetButton'; -import {WidgetDescription} from './widgetDescription'; -import {WidgetLayout} from './widgetLayout'; -import {WidgetTitle} from './widgetTitle'; - -<WidgetLayout - Title={<WidgetTitle title="epm() : /insights/frontend/assets" />} - Actions={ - <Fragment> - <WidgetButton>Say More</WidgetButton> - <WidgetButton>Say Less</WidgetButton> - <WidgetDescription - title="epm()" - description="Events received, tracked per minute" - /> - </Fragment> - } - Visualization={ - <LineChartWidgetVisualization timeseries={[sampleDurationTimeSeries]} /> - } - Footer={<span>This data is incomplete!</span>} -/> - - `} - </CodeSnippet> - - <SmallSizingWindow> - <WidgetLayout - Title={<WidgetTitle title="epm() : /insights/frontend/assets" />} - Actions={ - <Fragment> - <WidgetButton>Say More</WidgetButton> - <WidgetButton>Say Less</WidgetButton> - <WidgetDescription - title="epm() : /insights/frontend/assets" - description="Events received, tracked per minute" - /> - </Fragment> - } - Visualization={ - <LineChartWidgetVisualization timeSeries={[sampleDurationTimeSeries]} /> - } - Footer={<span>This data is incomplete!</span>} - /> - </SmallSizingWindow> - </Fragment> - ); - }); -}); - -const SmallSizingWindow = styled(SizingWindow)` - width: 400px; - height: 300px; -`; diff --git a/static/app/views/explore/charts/index.tsx b/static/app/views/explore/charts/index.tsx index cb3cbd5adcdc28..78f6ca1ee8725a 100644 --- a/static/app/views/explore/charts/index.tsx +++ b/static/app/views/explore/charts/index.tsx @@ -17,10 +17,7 @@ import usePageFilters from 'sentry/utils/usePageFilters'; import usePrevious from 'sentry/utils/usePrevious'; import {WidgetSyncContextProvider} from 'sentry/views/dashboards/contexts/widgetSyncContext'; import {TimeSeriesWidgetVisualization} from 'sentry/views/dashboards/widgets/timeSeriesWidget/timeSeriesWidgetVisualization'; -import {ErrorPanel} from 'sentry/views/dashboards/widgets/widgetLayout/errorPanel'; -import {LoadingPanel} from 'sentry/views/dashboards/widgets/widgetLayout/loadingPanel'; -import {WidgetLayout} from 'sentry/views/dashboards/widgets/widgetLayout/widgetLayout'; -import {WidgetTitle} from 'sentry/views/dashboards/widgets/widgetLayout/widgetTitle'; +import {Widget} from 'sentry/views/dashboards/widgets/widget/widget'; import {ConfidenceFooter} from 'sentry/views/explore/charts/confidenceFooter'; import ChartContextMenu from 'sentry/views/explore/components/chartContextMenu'; import { @@ -180,17 +177,19 @@ export function ExploreCharts({ const Title = ( <Fragment> {shouldRenderLabel && <ChartLabel>{chartInfo.label}</ChartLabel>} - <WidgetTitle title={chartInfo.formattedYAxes.filter(Boolean).join(', ')} /> + <Widget.WidgetTitle + title={chartInfo.formattedYAxes.filter(Boolean).join(', ')} + /> </Fragment> ); if (chartInfo.loading) { return ( - <WidgetLayout + <Widget key={index} height={CHART_HEIGHT} Title={Title} - Visualization={<LoadingPanel />} + Visualization={<TimeSeriesWidgetVisualization.LoadingPlaceholder />} revealActions="always" /> ); @@ -198,18 +197,18 @@ export function ExploreCharts({ if (chartInfo.error) { return ( - <WidgetLayout + <Widget key={index} height={CHART_HEIGHT} Title={Title} - Visualization={<ErrorPanel error={chartInfo.error} />} + Visualization={<Widget.WidgetError error={chartInfo.error} />} revealActions="always" /> ); } return ( - <WidgetLayout + <Widget key={index} height={CHART_HEIGHT} Title={Title}
0c25ab62e1a6e781727f5015760c764eb9e60d62
2024-05-07 23:56:29
Colleen O'Rourke
ref(rules): Don't fire delayed rules if they're snoozed (#70203)
false
Don't fire delayed rules if they're snoozed (#70203)
ref
diff --git a/src/sentry/rules/processing/delayed_processing.py b/src/sentry/rules/processing/delayed_processing.py index 556ee24b186f4a..038aa34ce28247 100644 --- a/src/sentry/rules/processing/delayed_processing.py +++ b/src/sentry/rules/processing/delayed_processing.py @@ -14,6 +14,7 @@ from sentry.models.grouprulestatus import GroupRuleStatus from sentry.models.project import Project from sentry.models.rule import Rule +from sentry.models.rulesnooze import RuleSnooze from sentry.rules import history, rules from sentry.rules.conditions.event_frequency import ( BaseEventFrequencyCondition, @@ -267,7 +268,10 @@ def apply_delayed(project_id: int) -> None: # STEP 3: Fetch the Rule models we need to check alert_rules = Rule.objects.filter(id__in=list(rules_to_groups.keys())) - + snoozed_rules = RuleSnooze.objects.filter(rule__in=alert_rules, user_id=None).values_list( + "rule", flat=True + ) + alert_rules = [rule for rule in alert_rules if rule.id not in snoozed_rules] # STEP 4: Create a map of unique conditions to a tuple containing the JSON # information needed to instantiate that condition class and the group_ids that # must be checked for that condition. We don't query per rule condition because @@ -286,7 +290,6 @@ def apply_delayed(project_id: int) -> None: ) # Step 7: Fire the rule's actions now = timezone.now() - # TODO: check rulesnooze table again before firing parsed_rulegroup_to_event_data = parse_rulegroup_to_event_data(rulegroup_to_event_data) for rule, group_ids in rules_to_fire.items(): diff --git a/tests/sentry/rules/processing/test_delayed_processing.py b/tests/sentry/rules/processing/test_delayed_processing.py index 21df01c6148771..979f18580ab35b 100644 --- a/tests/sentry/rules/processing/test_delayed_processing.py +++ b/tests/sentry/rules/processing/test_delayed_processing.py @@ -238,6 +238,34 @@ def test_apply_delayed_issue_platform_event(self): assert (self.rule1.id, self.group1.id) in rule_fire_histories assert (rule5.id, group5.id) in rule_fire_histories + def test_apply_delayed_snoozed_rule(self): + """ + Test that we do not fire a rule that's been snoozed (aka muted) + """ + rule5 = self.create_project_rule( + project=self.project, + condition_match=[self.event_frequency_condition2], + environment_id=self.environment.id, + ) + self.snooze_rule(owner_id=self.user.id, rule=rule5) + event5 = self.create_event(self.project, self.now, "group-5", self.environment.name) + self.create_event(self.project, self.now, "group-5", self.environment.name) + self.create_event(self.project, self.now, "group-5", self.environment.name) + group5 = event5.group + assert group5 + assert self.group1 + self.push_to_hash(self.project.id, rule5.id, group5.id, event5.event_id) + + with patch("sentry.buffer.backend.get_hash", self.redis_buffer.get_hash): + apply_delayed(self.project.id) + rule_fire_histories = RuleFireHistory.objects.filter( + rule__in=[rule5], + group__in=[self.group1, group5], + event_id__in=[self.event1.event_id, event5.event_id], + project=self.project, + ).values_list("rule", "group") + assert len(rule_fire_histories) == 0 + def test_apply_delayed_same_condition_diff_value(self): """ Test that two rules with the same condition and interval but a different value are both fired
fc35c436e5f3f076b5cf25a62b9c8e47760ef246
2024-02-22 01:21:16
Leander Rodrigues
ref(auto-assignment): Add logging to debug failed auto-assignments (#65555)
false
Add logging to debug failed auto-assignments (#65555)
ref
diff --git a/src/sentry/models/projectownership.py b/src/sentry/models/projectownership.py index 538f0217d15198..c0ff77e9062c91 100644 --- a/src/sentry/models/projectownership.py +++ b/src/sentry/models/projectownership.py @@ -111,11 +111,7 @@ def get_owners( For a given project_id, and event data blob. We combine the schemas from IssueOwners and CodeOwners. - If there are no matching rules, check ProjectOwnership.fallthrough: - If ProjectOwnership.fallthrough is enabled, return Everyone (all project members) - - we implicitly are falling through our rules and everyone is responsible. - If ProjectOwnership.fallthrough is disabled, return an empty list - - there are explicitly no owners + If there are no matching rules, return an empty list, and None for the rule. If there are matching rules, return the ordered actors. The order is determined by iterating through rules sequentially, evaluating @@ -269,26 +265,35 @@ def handle_auto_assignment( ownership = cls.get_ownership_cached(project_id) if not ownership: ownership = cls(project_id=project_id) + logging_extra["ownership"] = ownership autoassignment_types = cls._get_autoassignment_types(ownership) if not len(autoassignment_types): + logger.info("handle_auto_assignment.autoassignment_disabled", extra=logging_extra) return + logging_extra["autoassignment_types"] = autoassignment_types # Get the most recent GroupOwner that matches the following order: Suspect Committer, then Ownership Rule, then Code Owner issue_owner = GroupOwner.get_autoassigned_owner_cached( group.id, project_id, autoassignment_types ) if issue_owner is False: + logger.info("handle_auto_assignment.no_issue_owner", extra=logging_extra) return + logging_extra["issue_owner"] = issue_owner owner = issue_owner.owner() if not owner: + logger.info("handle_auto_assignment.no_owner", extra=logging_extra) return + logging_extra["owner"] = owner try: owner = owner.resolve() except (User.DoesNotExist, Team.DoesNotExist): + logger.info("handle_auto_assignment.no_resolved_owner", extra=logging_extra) return + logging_extra["resolved_owner"] = owner activity_details = {} if issue_owner.type == GroupOwnerType.SUSPECT_COMMIT.value: @@ -310,42 +315,55 @@ def handle_auto_assignment( if not auto_assigned and not force_autoassign: logger.info("autoassignment.post_manual_assignment", extra=logging_extra) return + + if not isinstance(owner, Team) and not isinstance(owner, RpcUser): + logging_extra["owner_type"] = str(type(owner)) + logger.info("handle_auto_assignment.unknown_owner_type", extra=logging_extra) + return + if ( isinstance(owner, Team) - and not GroupAssignee.objects.filter(group=group, team=owner.id).exists() - ) or ( + and GroupAssignee.objects.filter(group=group, team=owner.id).exists() + ): + logger.info("handle_auto_assignment.team_already_assigned", extra=logging_extra) + return + + if ( isinstance(owner, RpcUser) - and not GroupAssignee.objects.filter(group=group, user_id=owner.id).exists() + and GroupAssignee.objects.filter(group=group, user_id=owner.id).exists() ): - assignment = GroupAssignee.objects.assign( - group, - owner, - create_only=not force_autoassign, - extra=activity_details, - force_autoassign=force_autoassign, - ) + logger.info("handle_auto_assignment.user_already_assigned", extra=logging_extra) + return + + assignment = GroupAssignee.objects.assign( + group, + owner, + create_only=not force_autoassign, + extra=activity_details, + force_autoassign=force_autoassign, + ) - if assignment["new_assignment"] or assignment["updated_assignment"]: - analytics.record( - ( - "codeowners.assignment" - if activity_details.get("integration") - == ActivityIntegration.CODEOWNERS.value - else "issueowners.assignment" - ), - organization_id=organization_id or ownership.project.organization_id, - project_id=project_id, - group_id=group.id, - ) - logger.info( - "handle_auto_assignment.success", - extra={ - **logging_extra, - # owner_id returns a string including the owner type (user or team) and id - "assignee": issue_owner.owner_id(), - "reason": "created" if assignment["new_assignment"] else "updated", - }, - ) + if assignment["new_assignment"] or assignment["updated_assignment"]: + analytics.record( + ( + "codeowners.assignment" + if activity_details.get("integration") + == ActivityIntegration.CODEOWNERS.value + else "issueowners.assignment" + ), + organization_id=organization_id or ownership.project.organization_id, + project_id=project_id, + group_id=group.id, + ) + logger.info( + "handle_auto_assignment.success", + extra={ + **logging_extra, + # owner_id returns a string including the owner type (user or team) and id + "assignee": issue_owner.owner_id(), + "reason": "created" if assignment["new_assignment"] else "updated", + }, + ) @classmethod def _matching_ownership_rules(
7f01eefaa44cfabbf0d2202e61d7f62e7992f5b0
2022-11-02 19:27:30
Josh Ferge
feat(replays): add hasReplays to project serializer (#40863)
false
add hasReplays to project serializer (#40863)
feat
diff --git a/src/sentry/api/endpoints/organization_projects.py b/src/sentry/api/endpoints/organization_projects.py index a0edf3c73ca387..5810f335031eb7 100644 --- a/src/sentry/api/endpoints/organization_projects.py +++ b/src/sentry/api/endpoints/organization_projects.py @@ -72,6 +72,7 @@ class OrganizationProjectsEndpoint(OrganizationEndpoint, EnvironmentMixin): "firstTransactionEvent": True, "hasSessions": True, "hasProfiles": True, + "hasReplays": True, "latestRelease": None, "hasUserReports": False, } diff --git a/src/sentry/api/serializers/models/project.py b/src/sentry/api/serializers/models/project.py index 6f3f676fa4dd92..3ed6ab4cbb49bc 100644 --- a/src/sentry/api/serializers/models/project.py +++ b/src/sentry/api/serializers/models/project.py @@ -206,6 +206,7 @@ class ProjectSerializerBaseResponse(_ProjectSerializerOptionalBaseResponse): firstTransactionEvent: bool hasSessions: bool hasProfiles: bool + hasReplays: bool platform: Optional[str] firstEvent: Optional[datetime] @@ -463,6 +464,7 @@ def serialize(self, obj, attrs, user) -> ProjectSerializerResponse: "firstTransactionEvent": bool(obj.flags.has_transactions), "hasSessions": bool(obj.flags.has_sessions), "hasProfiles": bool(obj.flags.has_profiles), + "hasReplays": bool(obj.flags.has_replays), "features": attrs["features"], "status": status_label, "platform": obj.platform, @@ -687,6 +689,7 @@ def serialize(self, obj, attrs, user) -> OrganizationProjectResponse: # type: i firstTransactionEvent=bool(obj.flags.has_transactions), hasSessions=bool(obj.flags.has_sessions), hasProfiles=bool(obj.flags.has_profiles), + hasReplays=bool(obj.flags.has_replays), platform=obj.platform, platforms=attrs["platforms"], latestRelease=attrs["latest_release"], diff --git a/tests/sentry/api/serializers/test_project.py b/tests/sentry/api/serializers/test_project.py index 735089e586d604..024e4d6a2036f0 100644 --- a/tests/sentry/api/serializers/test_project.py +++ b/tests/sentry/api/serializers/test_project.py @@ -328,6 +328,16 @@ def test_has_profiles_flag(self): result = serialize(self.project, self.user, ProjectSummarySerializer()) assert result["hasProfiles"] is True + def test_has_replays_flag(self): + result = serialize(self.project, self.user, ProjectSummarySerializer()) + assert result["hasReplays"] is False + + self.project.first_event = timezone.now() + self.project.update(flags=F("flags").bitor(Project.flags.has_replays)) + + result = serialize(self.project, self.user, ProjectSummarySerializer()) + assert result["hasReplays"] is True + def test_no_environments(self): # remove environments and related models Deploy.objects.all().delete() @@ -494,7 +504,7 @@ def test_stats_with_sessions( assert results[0]["sessionStats"]["currentCrashFreeRate"] == 75.63453 assert results[0]["sessionStats"]["hasHealthData"] - check_has_health_data.assert_not_called() + assert check_has_health_data.call_count == 0 @mock.patch("sentry.api.serializers.models.project.release_health.check_has_health_data") @mock.patch( @@ -523,7 +533,7 @@ def test_stats_with_sessions_and_none_crash_free_rates( assert results[0]["sessionStats"]["currentCrashFreeRate"] is None assert results[0]["sessionStats"]["hasHealthData"] - check_has_health_data.assert_called() + assert check_has_health_data.call_count == 1 @region_silo_test
0e5172e5af9b5981c740987171daf4578b3efb3e
2018-03-21 00:31:03
Lyn Nagara
fix(environments): Prevent double API request when environment is changed (#7714)
false
Prevent double API request when environment is changed (#7714)
fix
diff --git a/src/sentry/static/sentry/app/views/groupUserReports.jsx b/src/sentry/static/sentry/app/views/groupUserReports.jsx index 12eb2972b403ff..d1abf9375ba112 100644 --- a/src/sentry/static/sentry/app/views/groupUserReports.jsx +++ b/src/sentry/static/sentry/app/views/groupUserReports.jsx @@ -1,6 +1,9 @@ import React from 'react'; import createReactClass from 'create-react-class'; import {Link} from 'react-router'; +import {omit, isEqual} from 'lodash'; +import qs from 'query-string'; + import SentryTypes from '../proptypes'; import ApiMixin from '../mixins/apiMixin'; import GroupState from '../mixins/groupState'; @@ -33,10 +36,14 @@ const GroupUserReports = createReactClass({ }, componentDidUpdate(prevProps) { - if ( - prevProps.location.search !== this.props.location.search || - prevProps.environment !== this.props.environment - ) { + // Search term has changed (excluding environment) + const searchHasChanged = !isEqual( + omit(qs.parse(prevProps.location.search), 'environment'), + omit(qs.parse(this.props.location.search), 'environment') + ); + const environmentHasChanged = prevProps.environment !== this.props.environment; + + if (searchHasChanged || environmentHasChanged) { this.fetchData(); } },
41463266544fbe1bf62484a38e4636f889f75ae5
2021-06-14 17:28:15
Ahmed Etefy
feat(api): Adds Crash free rate + `hasHealthData` check to `OrganizationProjectsEndpoint` (#26586)
false
Adds Crash free rate + `hasHealthData` check to `OrganizationProjectsEndpoint` (#26586)
feat
diff --git a/src/sentry/api/endpoints/organization_projects.py b/src/sentry/api/endpoints/organization_projects.py index 6e6288695d47b8..d9363f30db7df9 100644 --- a/src/sentry/api/endpoints/organization_projects.py +++ b/src/sentry/api/endpoints/organization_projects.py @@ -102,11 +102,13 @@ def get(self, request, organization): def serialize_on_result(result): transaction_stats = request.GET.get("transactionStats") + session_stats = request.GET.get("sessionStats") environment_id = self._get_environment_id_from_request(request, organization.id) serializer = ProjectSummarySerializer( environment_id=environment_id, stats_period=stats_period, transaction_stats=transaction_stats, + session_stats=session_stats, collapse=collapse, ) return serialize(result, request.user, serializer) diff --git a/src/sentry/api/serializers/models/project.py b/src/sentry/api/serializers/models/project.py index 5fa28f0d864a22..3e2b83b3800cb7 100644 --- a/src/sentry/api/serializers/models/project.py +++ b/src/sentry/api/serializers/models/project.py @@ -36,6 +36,7 @@ from sentry.notifications.helpers import transform_to_notification_settings_by_parent_id from sentry.notifications.types import NotificationSettingOptionValues, NotificationSettingTypes from sentry.snuba import discover +from sentry.snuba.sessions import check_has_health_data, get_current_and_previous_crash_free_rates from sentry.types.integrations import ExternalProviders from sentry.utils.compat import zip @@ -155,6 +156,7 @@ def __init__( environment_id: Optional[str] = None, stats_period: Optional[str] = None, transaction_stats: Optional[str] = None, + session_stats: Optional[str] = None, ) -> None: if stats_period is not None: assert stats_period in STATS_PERIOD_CHOICES @@ -162,6 +164,7 @@ def __init__( self.environment_id = environment_id self.stats_period = stats_period self.transaction_stats = transaction_stats + self.session_stats = session_stats def get_attrs( self, item_list: Sequence[Project], user: User, **kwargs: Any @@ -201,12 +204,15 @@ def measure_span(op_tag): with measure_span("stats"): stats = None transaction_stats = None + session_stats = None project_ids = [o.id for o in item_list] if self.transaction_stats and self.stats_period: stats = self.get_stats(project_ids, "!event.type:transaction") transaction_stats = self.get_stats(project_ids, "event.type:transaction") elif self.stats_period: stats = self.get_stats(project_ids, "!event.type:transaction") + if self.session_stats: + session_stats = self.get_session_stats(project_ids) avatars = {a.project_id: a for a in ProjectAvatar.objects.filter(project__in=item_list)} project_ids = [i.id for i in item_list] @@ -243,6 +249,8 @@ def measure_span(op_tag): serialized["stats"] = stats[project.id] if transaction_stats: serialized["transactionStats"] = transaction_stats[project.id] + if session_stats: + serialized["sessionStats"] = session_stats[project.id] return result def get_stats(self, project_ids, query): @@ -282,6 +290,47 @@ def get_stats(self, project_ids, query): results[project_id] = serialized return results + def get_session_stats(self, project_ids): + segments, interval = STATS_PERIOD_CHOICES[self.stats_period] + + now = timezone.now() + current_interval_start = now - (segments * interval) + previous_interval_start = now - (2 * segments * interval) + + project_health_data_dict = get_current_and_previous_crash_free_rates( + project_ids=project_ids, + current_start=current_interval_start, + current_end=now, + previous_start=previous_interval_start, + previous_end=current_interval_start, + rollup=int(interval.total_seconds()), + ) + + # list that contains ids of projects that has both `currentCrashFreeRate` and + # `previousCrashFreeRate` set to None and so we are not sure if they have health data or + # not and so we add those ids to this list to check later + check_has_health_data_ids = [] + + for project_id in project_ids: + current_crash_free_rate = project_health_data_dict[project_id]["currentCrashFreeRate"] + previous_crash_free_rate = project_health_data_dict[project_id]["previousCrashFreeRate"] + + if [current_crash_free_rate, previous_crash_free_rate] != [None, None]: + project_health_data_dict[project_id]["hasHealthData"] = True + else: + project_health_data_dict[project_id]["hasHealthData"] = False + check_has_health_data_ids.append(project_id) + + # For project ids we are not sure if they have health data in the last 90 days we + # call -> check_has_data with those ids and then update our `project_health_data_dict` + # accordingly + if check_has_health_data_ids: + projects_with_health_data = check_has_health_data(check_has_health_data_ids) + for project_id in projects_with_health_data: + project_health_data_dict[project_id]["hasHealthData"] = True + + return project_health_data_dict + def serialize(self, obj, attrs, user): status_label = STATUS_LABELS.get(obj.status, "unknown") @@ -315,6 +364,8 @@ def serialize(self, obj, attrs, user): context["stats"] = attrs["stats"] if "transactionStats" in attrs: context["transactionStats"] = attrs["transactionStats"] + if "sessionStats" in attrs: + context["sessionStats"] = attrs["sessionStats"] return context @@ -375,12 +426,15 @@ def serialize(self, obj, attrs, user): class ProjectSummarySerializer(ProjectWithTeamSerializer): def __init__( - self, environment_id=None, stats_period=None, transaction_stats=None, collapse=None + self, + environment_id=None, + stats_period=None, + transaction_stats=None, + session_stats=None, + collapse=None, ): super(ProjectWithTeamSerializer, self).__init__( - environment_id, - stats_period, - transaction_stats, + environment_id, stats_period, transaction_stats, session_stats ) self.collapse = collapse @@ -509,6 +563,8 @@ def serialize(self, obj, attrs, user): context["stats"] = attrs["stats"] if "transactionStats" in attrs: context["transactionStats"] = attrs["transactionStats"] + if "sessionStats" in attrs: + context["sessionStats"] = attrs["sessionStats"] return context diff --git a/src/sentry/snuba/sessions.py b/src/sentry/snuba/sessions.py index fa290cd287ed46..1b7ac3de64b9a5 100644 --- a/src/sentry/snuba/sessions.py +++ b/src/sentry/snuba/sessions.py @@ -101,21 +101,49 @@ def get_oldest_health_data_for_releases(project_releases): return rv -def check_has_health_data(project_releases): - conditions = [["release", "IN", list(x[1] for x in project_releases)]] - filter_keys = {"project_id": list({x[0] for x in project_releases})} - return { - (x["project_id"], x["release"]) - for x in raw_query( - dataset=Dataset.Sessions, - selected_columns=["release", "project_id"], - groupby=["release", "project_id"], - start=datetime.utcnow() - timedelta(days=90), - conditions=conditions, - referrer="sessions.health-data-check", - filter_keys=filter_keys, - )["data"] +def check_has_health_data(projects_list): + """ + Function that returns a set of all project_ids or (project, release) if they have health data + within the last 90 days based on a list of projects or a list of project, release combinations + provided as an arg. + Inputs: + * projects_list: Contains either a list of project ids or a list of tuple (project_id, + release) + """ + if len(projects_list) == 0: + return set() + + conditions = None + # Check if projects_list also contains releases as a tuple of (project_id, releases) + includes_releases = type(projects_list[0]) == tuple + + if includes_releases: + filter_keys = {"project_id": list({x[0] for x in projects_list})} + conditions = [["release", "IN", list(x[1] for x in projects_list)]] + query_cols = ["release", "project_id"] + + def data_tuple(x): + return x["project_id"], x["release"] + + else: + filter_keys = {"project_id": list({x for x in projects_list})} + query_cols = ["project_id"] + + def data_tuple(x): + return x["project_id"] + + raw_query_args = { + "dataset": Dataset.Sessions, + "selected_columns": query_cols, + "groupby": query_cols, + "start": datetime.utcnow() - timedelta(days=90), + "referrer": "sessions.health-data-check", + "filter_keys": filter_keys, } + if conditions is not None: + raw_query_args.update({"conditions": conditions}) + + return {data_tuple(x) for x in raw_query(**raw_query_args)["data"]} def get_project_releases_by_stability( @@ -916,3 +944,90 @@ def __get_scope_value_for_release( elif scope == "crash_free_users": scope_value = rq_row["users_crashed"] / rq_row["users"] return scope_value + + +def __get_crash_free_rate_data(project_ids, start, end, rollup): + """ + Helper function that executes a snuba query on project_ids to fetch the number of crashed + sessions and total sessions and returns the crash free rate for those project_ids. + Inputs: + * project_ids + * start + * end + * rollup + Returns: + Snuba query results + """ + return raw_query( + dataset=Dataset.Sessions, + selected_columns=["project_id", "sessions_crashed", "sessions"], + filter_keys={"project_id": project_ids}, + start=start, + end=end, + rollup=rollup, + groupby=["project_id"], + referrer="sessions.totals", + )["data"] + + +def get_current_and_previous_crash_free_rates( + project_ids, current_start, current_end, previous_start, previous_end, rollup +): + """ + Function that returns `currentCrashFreeRate` and the `previousCrashFreeRate` of projects + based on the inputs provided + Inputs: + * project_ids + * current_start: start interval of currentCrashFreeRate + * current_end: end interval of currentCrashFreeRate + * previous_start: start interval of previousCrashFreeRate + * previous_end: end interval of previousCrashFreeRate + * rollup + Returns: + A dictionary of project_id as key and as value the `currentCrashFreeRate` and the + `previousCrashFreeRate` + + As an example: + { + 1: { + "currentCrashFreeRate": 100, + "previousCrashFreeRate": 66.66666666666667 + }, + 2: { + "currentCrashFreeRate": 50.0, + "previousCrashFreeRate": None + }, + ... + } + """ + projects_crash_free_rate_dict = { + prj: {"currentCrashFreeRate": None, "previousCrashFreeRate": None} for prj in project_ids + } + + def calculate_crash_free_percentage(row): + return 100 - (row["sessions_crashed"] / row["sessions"]) * 100 + + # currentCrashFreeRate + current_crash_free_data = __get_crash_free_rate_data( + project_ids=project_ids, + start=current_start, + end=current_end, + rollup=rollup, + ) + for row in current_crash_free_data: + projects_crash_free_rate_dict[row["project_id"]].update( + {"currentCrashFreeRate": calculate_crash_free_percentage(row)} + ) + + # previousCrashFreeRate + previous_crash_free_data = __get_crash_free_rate_data( + project_ids=project_ids, + start=previous_start, + end=previous_end, + rollup=rollup, + ) + for row in previous_crash_free_data: + projects_crash_free_rate_dict[row["project_id"]].update( + {"previousCrashFreeRate": calculate_crash_free_percentage(row)} + ) + return projects_crash_free_rate_dict diff --git a/tests/sentry/api/serializers/test_project.py b/tests/sentry/api/serializers/test_project.py index 16b411b0137878..26fe644f31c4e6 100644 --- a/tests/sentry/api/serializers/test_project.py +++ b/tests/sentry/api/serializers/test_project.py @@ -442,6 +442,54 @@ def test_stats_with_transactions(self): assert 24 == len(results[0]["transactionStats"]) assert [1] == [v[1] for v in results[0]["transactionStats"] if v[1] > 0] + @mock.patch("sentry.api.serializers.models.project.check_has_health_data") + @mock.patch("sentry.api.serializers.models.project.get_current_and_previous_crash_free_rates") + def test_stats_with_sessions( + self, get_current_and_previous_crash_free_rates, check_has_health_data + ): + get_current_and_previous_crash_free_rates.return_value = { + self.project.id: { + "currentCrashFreeRate": 75.63453, + "previousCrashFreeRate": 99.324543, + } + } + serializer = ProjectSummarySerializer(stats_period="24h", session_stats=True) + results = serialize([self.project], self.user, serializer) + + assert "sessionStats" in results[0] + assert results[0]["sessionStats"]["previousCrashFreeRate"] == 99.324543 + assert results[0]["sessionStats"]["currentCrashFreeRate"] == 75.63453 + assert results[0]["sessionStats"]["hasHealthData"] + + check_has_health_data.assert_not_called() # NOQA + + @mock.patch("sentry.api.serializers.models.project.check_has_health_data") + @mock.patch("sentry.api.serializers.models.project.get_current_and_previous_crash_free_rates") + def test_stats_with_sessions_and_none_crash_free_rates( + self, get_current_and_previous_crash_free_rates, check_has_health_data + ): + """ + Test that ensures if both `currentCrashFreeRate` and `previousCrashFreeRate` are None, then + we need to make a call to `check_has_health_data` to know if we have health data in that + specific project_id(s) + """ + check_has_health_data.return_value = {self.project.id} + get_current_and_previous_crash_free_rates.return_value = { + self.project.id: { + "currentCrashFreeRate": None, + "previousCrashFreeRate": None, + } + } + serializer = ProjectSummarySerializer(stats_period="24h", session_stats=True) + results = serialize([self.project], self.user, serializer) + + assert "sessionStats" in results[0] + assert results[0]["sessionStats"]["previousCrashFreeRate"] is None + assert results[0]["sessionStats"]["currentCrashFreeRate"] is None + assert results[0]["sessionStats"]["hasHealthData"] + + check_has_health_data.assert_called() # NOQA + class ProjectWithOrganizationSerializerTest(TestCase): def test_simple(self): diff --git a/tests/snuba/sessions/test_sessions.py b/tests/snuba/sessions/test_sessions.py index c632481fc9bc97..a8fcdaa7ea7f33 100644 --- a/tests/snuba/sessions/test_sessions.py +++ b/tests/snuba/sessions/test_sessions.py @@ -1,13 +1,15 @@ import time import uuid -from datetime import datetime +from datetime import datetime, timedelta import pytz +from django.utils import timezone from sentry.snuba.sessions import ( _make_stats, check_has_health_data, get_adjacent_releases_based_on_adoption, + get_current_and_previous_crash_free_rates, get_oldest_health_data_for_releases, get_project_releases_by_stability, get_release_adoption, @@ -15,6 +17,7 @@ get_release_sessions_time_bounds, ) from sentry.testutils import SnubaTestCase, TestCase +from sentry.utils.dates import to_timestamp def format_timestamp(dt): @@ -27,6 +30,26 @@ def make_24h_stats(ts): return _make_stats(datetime.utcfromtimestamp(ts).replace(tzinfo=pytz.utc), 3600, 24) +def generate_session_default_args(session_dict): + session_dict_default = { + "session_id": str(uuid.uuid4()), + "distinct_id": str(uuid.uuid4()), + "status": "ok", + "seq": 0, + "release": "[email protected]", + "environment": "prod", + "retention_days": 90, + "org_id": 0, + "project_id": 0, + "duration": 60.0, + "errors": 0, + "started": time.time() // 60 * 60, + "received": time.time(), + } + session_dict_default.update(session_dict) + return session_dict_default + + class SnubaSessionsTest(TestCase, SnubaTestCase): def setUp(self): super().setUp() @@ -120,6 +143,58 @@ def test_check_has_health_data(self): ) assert data == {(self.project.id, self.session_release)} + def test_check_has_health_data_without_releases_should_exlude_sessions_gt_90_days(self): + """ + Test that ensures that `check_has_health_data` returns a set of projects that has health + data within the last 90d if only a list of project ids is provided and that any project + with session data older than 90 days should be exluded + """ + project2 = self.create_project( + name="Bar2", + slug="bar2", + teams=[self.team], + fire_project_created=True, + organization=self.organization, + ) + + date_100_days_ago = to_timestamp( + (datetime.utcnow() - timedelta(days=100)).replace(tzinfo=pytz.utc) + ) + self.store_session( + generate_session_default_args( + { + "started": date_100_days_ago // 60 * 60, + "received": date_100_days_ago, + "project_id": project2.id, + "org_id": project2.organization_id, + "status": "exited", + } + ) + ) + data = check_has_health_data([self.project.id, project2.id]) + assert data == {self.project.id} + + def test_check_has_health_data_without_releases_should_include_sessions_lte_90_days(self): + """ + Test that ensures that `check_has_health_data` returns a set of projects that has health + data within the last 90d if only a list of project ids is provided and any project with + session data earlier than 90 days should be included + """ + project2 = self.create_project( + name="Bar2", + slug="bar2", + teams=[self.team], + fire_project_created=True, + organization=self.organization, + ) + self.store_session( + generate_session_default_args( + {"project_id": project2.id, "org_id": project2.organization_id, "status": "exited"} + ) + ) + data = check_has_health_data([self.project.id, project2.id]) + assert data == {self.project.id, project2.id} + def test_get_project_releases_by_stability(self): # Add an extra session with a different `distinct_id` so that sorting by users # is stable @@ -456,26 +531,6 @@ def compare_releases_list_according_to_current_release_filters( adjacent_releases = get_adjacent_releases_based_on_adoption(**adj_releases_filters) assert adjacent_releases == releases_list - @staticmethod - def generate_session_default_args(session_dict): - session_dict_default = { - "session_id": str(uuid.uuid4()), - "distinct_id": str(uuid.uuid4()), - "status": "ok", - "seq": 0, - "release": "[email protected]", - "environment": "prod", - "retention_days": 90, - "org_id": 0, - "project_id": 0, - "duration": 60.0, - "errors": 0, - "started": time.time() // 60 * 60, - "received": time.time(), - } - session_dict_default.update(session_dict) - return session_dict_default - class SnubaReleaseDetailPaginationOnSessionsTest( TestCase, SnubaTestCase, SnubaReleaseDetailPaginationBaseTestClass @@ -527,7 +582,7 @@ def setUp(self): # Time: < 24h # Total: 1 Session self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -542,7 +597,7 @@ def setUp(self): # Total: 2 sessions for _ in range(0, 2): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started_gt_24h, @@ -556,7 +611,7 @@ def setUp(self): # Time: < 24h # Total: 1 Session self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -572,7 +627,7 @@ def setUp(self): # Time: < 24h # Total: 2 Sessions self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -584,7 +639,7 @@ def setUp(self): ) ) self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -597,7 +652,7 @@ def setUp(self): ) ) self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -613,7 +668,7 @@ def setUp(self): # Total: 3 Session for _ in range(0, 3): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -629,7 +684,7 @@ def setUp(self): # Total: 3 Sessions for _ in range(0, 3): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -788,7 +843,7 @@ def setUp(self): # Time: < 24h # Total: 1 Session -> 100% Crash Free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -803,7 +858,7 @@ def setUp(self): # Total: 3 sessions -> 1 Healthy + 2 Crashed -> 33.3333% Crash Free for _ in range(0, 2): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started_gt_24h, @@ -817,7 +872,7 @@ def setUp(self): # Time: < 24h # Total: 1 Session -> 0% Crash Free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -833,7 +888,7 @@ def setUp(self): # Time: < 24h # Total: 2 Sessions -> 2 Crashed -> 0% Crash free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -845,7 +900,7 @@ def setUp(self): ) ) self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -858,7 +913,7 @@ def setUp(self): ) ) self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -873,7 +928,7 @@ def setUp(self): # Time: <24h # Total: 3 Session -> 2 Healthy + 1 Crashed -> 66.666% Crash free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -884,7 +939,7 @@ def setUp(self): ) for _ in range(0, 2): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -899,7 +954,7 @@ def setUp(self): # Time: <24h # Total: 3 Sessions -> 2 Healthy + 1 Crashed -> 66.666% Crash free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -910,7 +965,7 @@ def setUp(self): ) for _ in range(0, 2): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1060,7 +1115,7 @@ def setUp(self): # Time: < 24h # Total: 1 Session self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1075,7 +1130,7 @@ def setUp(self): # Total: 2 sessions for _ in range(0, 2): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started_gt_24h, @@ -1089,7 +1144,7 @@ def setUp(self): # Time: < 24h # Total: 1 Session self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1105,7 +1160,7 @@ def setUp(self): # Time: < 24h # Total: 2 Sessions self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1117,7 +1172,7 @@ def setUp(self): ) ) self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1131,7 +1186,7 @@ def setUp(self): ) self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1147,7 +1202,7 @@ def setUp(self): # Total: 3 Session for _ in range(0, 3): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1163,7 +1218,7 @@ def setUp(self): # Total: 3 Sessions for _ in range(0, 3): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1313,7 +1368,7 @@ def setUp(self): # Time: < 24h # Total: 1 Session -> 100% Crash Free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1328,7 +1383,7 @@ def setUp(self): # Total: 3 sessions -> 1 Healthy + 2 Crashed -> 33.3333% Crash Free for _ in range(0, 2): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started_gt_24h, @@ -1341,7 +1396,7 @@ def setUp(self): # Time: < 24h # Total: 1 Session -> 0% Crash Free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1357,7 +1412,7 @@ def setUp(self): # Time: < 24h # Total: 2 Sessions -> 2 Crashed -> 0% Crash free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1369,7 +1424,7 @@ def setUp(self): ) ) self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1382,7 +1437,7 @@ def setUp(self): ) ) self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1397,7 +1452,7 @@ def setUp(self): # Time: <24h # Total: 3 Session -> 2 Healthy + 1 Crashed -> 66.666% Crash free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1408,7 +1463,7 @@ def setUp(self): ) for _ in range(0, 2): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1423,7 +1478,7 @@ def setUp(self): # Time: <24h # Total: 3 Sessions -> 2 Healthy + 1 Crashed -> 66.666% Crash free self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1434,7 +1489,7 @@ def setUp(self): ) for _ in range(0, 2): self.store_session( - self.generate_session_default_args( + generate_session_default_args( { **self.common_session_args, "started": self.session_started, @@ -1533,3 +1588,123 @@ def test_get_adjacent_releases_to_last_release_in_different_env( "prev_releases_list": [], }, ) + + +class GetCrashFreeRateTestCase(TestCase, SnubaTestCase): + """ + TestClass that tests that `get_current_and_previous_crash_free_rates` returns the correct + `currentCrashFreeRate` and `previousCrashFreeRate` for each project + + TestData: + Project 1: + In the last 24h -> 2 Exited Sessions / 2 Total Sessions -> 100% Crash free rate + In the previous 24h (>24h & <48h) -> 2 Exited + 1 Crashed Sessions / 3 Sessions -> 66.7% + + Project 2: + In the last 24h -> 1 Exited + 1 Crashed / 2 Total Sessions -> 50% Crash free rate + In the previous 24h (>24h & <48h) -> 0 Sessions -> None + + Project 3: + In the last 24h -> 0 Sessions -> None + In the previous 24h (>24h & <48h) -> 4 Exited + 1 Crashed / 5 Total Sessions -> 80% + """ + + def setUp(self): + super().setUp() + self.session_started = time.time() // 60 * 60 + self.session_started_gt_24_lt_48 = self.session_started - 30 * 60 * 60 + self.project2 = self.create_project( + name="Bar2", + slug="bar2", + teams=[self.team], + fire_project_created=True, + organization=self.organization, + ) + self.project3 = self.create_project( + name="Bar3", + slug="bar3", + teams=[self.team], + fire_project_created=True, + organization=self.organization, + ) + + # Project 1 + for _ in range(0, 2): + self.store_session( + generate_session_default_args( + { + "project_id": self.project.id, + "org_id": self.project.organization_id, + "status": "exited", + } + ) + ) + + for idx in range(0, 3): + status = "exited" + if idx == 2: + status = "crashed" + self.store_session( + generate_session_default_args( + { + "project_id": self.project.id, + "org_id": self.project.organization_id, + "status": status, + "started": self.session_started_gt_24_lt_48, + } + ) + ) + + # Project 2 + for i in range(0, 2): + status = "exited" + if i == 1: + status = "crashed" + self.store_session( + generate_session_default_args( + { + "project_id": self.project2.id, + "org_id": self.project2.organization_id, + "status": status, + } + ) + ) + + # Project 3 + for i in range(0, 5): + status = "exited" + if i == 4: + status = "crashed" + self.store_session( + generate_session_default_args( + { + "project_id": self.project3.id, + "org_id": self.project3.organization_id, + "status": status, + "started": self.session_started_gt_24_lt_48, + } + ) + ) + + def test_get_current_and_previous_crash_free_rates(self): + now = timezone.now() + last_24h_start = now - 24 * timedelta(hours=1) + last_48h_start = now - 2 * 24 * timedelta(hours=1) + + data = get_current_and_previous_crash_free_rates( + project_ids=[self.project.id, self.project2.id, self.project3.id], + current_start=last_24h_start, + current_end=now, + previous_start=last_48h_start, + previous_end=last_24h_start, + rollup=86400, + ) + + assert data == { + self.project.id: { + "currentCrashFreeRate": 100, + "previousCrashFreeRate": 66.66666666666667, + }, + self.project2.id: {"currentCrashFreeRate": 50.0, "previousCrashFreeRate": None}, + self.project3.id: {"currentCrashFreeRate": None, "previousCrashFreeRate": 80.0}, + }
1a0f8ac51b8204ac6735bbe35b28638e2568bd80
2018-02-16 22:11:22
Max Bittker
feat(subscribe): Bulk subscribe and subscribe_actor (#7261)
false
Bulk subscribe and subscribe_actor (#7261)
feat
diff --git a/src/sentry/api/endpoints/group_notes.py b/src/sentry/api/endpoints/group_notes.py index 8631653342c2d9..7f3a075c2332cc 100644 --- a/src/sentry/api/endpoints/group_notes.py +++ b/src/sentry/api/endpoints/group_notes.py @@ -73,19 +73,21 @@ def post(self, request, group): mentioned_teams = actor_mentions.get('teams') - mentioned_team_users = User.objects.filter( - sentry_orgmember_set__organization_id=group.project.organization_id, - sentry_orgmember_set__organizationmemberteam__team__in=mentioned_teams, - sentry_orgmember_set__organizationmemberteam__is_active=True, - is_active=True, - ).exclude(id__in={u.id for u in actor_mentions.get('users')}) - - for user in mentioned_team_users: - GroupSubscription.objects.subscribe( - group=group, - user=user, - reason=GroupSubscriptionReason.team_mentioned, - ) + mentioned_team_users = list( + User.objects.filter( + sentry_orgmember_set__organization_id=group.project.organization_id, + sentry_orgmember_set__organizationmemberteam__team__in=mentioned_teams, + sentry_orgmember_set__organizationmemberteam__is_active=True, + is_active=True, + ).exclude(id__in={u.id for u in actor_mentions.get('users')}) + .values_list('id', flat=True) + ) + + GroupSubscription.objects.bulk_subscribe( + group=group, + user_ids=mentioned_team_users, + reason=GroupSubscriptionReason.team_mentioned, + ) activity = Activity.objects.create( group=group, diff --git a/src/sentry/models/groupsubscription.py b/src/sentry/models/groupsubscription.py index 9b5271b7984344..8af92722b603f4 100644 --- a/src/sentry/models/groupsubscription.py +++ b/src/sentry/models/groupsubscription.py @@ -94,6 +94,54 @@ def subscribe(self, group, user, reason=GroupSubscriptionReason.unknown): except IntegrityError: pass + def subscribe_actor(self, group, actor, reason=GroupSubscriptionReason.unknown): + from sentry.models import User, Team + + if isinstance(actor, User): + return self.subscribe(group, actor, reason) + if isinstance(actor, Team): + # subscribe the members of the team + team_users_ids = list(actor.member_set.values_list('user_id', flat=True)) + return self.bulk_subscribe(group, team_users_ids, reason) + + raise NotImplementedError('Unknown actor type: %r' % type(actor)) + + def bulk_subscribe(self, group, user_ids, reason=GroupSubscriptionReason.unknown): + """ + Subscribe a list of user ids to an issue, but only if the users are not explicitly + unsubscribed. + """ + # 5 retries for race conditions where + # concurrent subscription attempts cause integrity errors + for _ in range(5): + + existing_subscriptions = set(GroupSubscription.objects.filter( + user_id__in=user_ids, + group=group, + project=group.project, + ).values_list('user_id', flat=True)) + + subscriptions = [ + GroupSubscription( + user_id=user_id, + group=group, + project=group.project, + is_active=True, + reason=reason, + ) + for user_id in user_ids + if user_id not in existing_subscriptions + ] + + try: + with transaction.atomic(): + self.bulk_create(subscriptions) + return True + except IntegrityError: + pass + + raise Exception('Bulk_Subscribe failed') + def get_participants(self, group): """ Identify all users who are participating with a given issue. diff --git a/tests/sentry/models/test_groupsubscription.py b/tests/sentry/models/test_groupsubscription.py index 157c8e0be06b4b..08a7d71a922815 100644 --- a/tests/sentry/models/test_groupsubscription.py +++ b/tests/sentry/models/test_groupsubscription.py @@ -24,6 +24,67 @@ def test_simple(self): # should not error GroupSubscription.objects.subscribe(group=group, user=user) + def test_bulk(self): + group = self.create_group() + + user_ids = [] + for i in range(20): + user = self.create_user() + user_ids.append(user.id) + + GroupSubscription.objects.bulk_subscribe(group=group, user_ids=user_ids) + + assert len(GroupSubscription.objects.filter( + group=group, + )) == 20 + + one_more = self.create_user() + user_ids.append(one_more.id) + + # should not error + GroupSubscription.objects.bulk_subscribe(group=group, user_ids=user_ids) + + assert len(GroupSubscription.objects.filter( + group=group, + )) == 21 + + def test_actor_user(self): + group = self.create_group() + user = self.create_user() + + GroupSubscription.objects.subscribe_actor(group=group, actor=user) + + assert GroupSubscription.objects.filter( + group=group, + user=user, + ).exists() + + # should not error + GroupSubscription.objects.subscribe_actor(group=group, actor=user) + + def test_actor_team(self): + org = self.create_organization() + group = self.create_group(organization=org) + user = self.create_user() + team = self.create_team(organization=org) + self.create_member( + user=user, + email='[email protected]', + organization=org, + role='owner', + teams=[team], + ) + + GroupSubscription.objects.subscribe_actor(group=group, actor=team) + + assert GroupSubscription.objects.filter( + group=group, + user=user, + ).exists() + + # should not error + GroupSubscription.objects.subscribe_actor(group=group, actor=team) + class GetParticipantsTest(TestCase): def test_simple(self):
306fa43d7252836d0011e4da2699e6adafcf03b2
2024-07-31 18:37:32
Ogi
feat(minimetrics): attach tags as data (#75324)
false
attach tags as data (#75324)
feat
diff --git a/src/sentry/metrics/minimetrics.py b/src/sentry/metrics/minimetrics.py index 5f005beed9dbc5..446b878d002bbb 100644 --- a/src/sentry/metrics/minimetrics.py +++ b/src/sentry/metrics/minimetrics.py @@ -8,10 +8,10 @@ from sentry.metrics.base import MetricsBackend, Tags -def _set_tags(span: Span, tags: Tags | None) -> None: +def _attach_tags(span: Span, tags: Tags | None) -> None: if tags: for tag_key, tag_value in tags.items(): - span.set_tag(tag_key, tag_value) + span.set_data(tag_key, tag_value) @metrics_noop @@ -21,7 +21,7 @@ def _set_metric_on_span(key: str, value: float | int, op: str, tags: Tags | None return span_or_tx.set_data(key, value) - _set_tags(span_or_tx, tags) + _attach_tags(span_or_tx, tags) class MiniMetricsMetricsBackend(MetricsBackend): @@ -57,13 +57,13 @@ def timing( return if span_or_tx.op == key: - _set_tags(span_or_tx, tags) + _attach_tags(span_or_tx, tags) return timestamp = datetime.now(timezone.utc) start_timestamp = timestamp - timedelta(seconds=value) span = span_or_tx.start_child(op=key, start_timestamp=start_timestamp) - _set_tags(span, tags) + _attach_tags(span, tags) span.finish(end_timestamp=timestamp) def gauge( diff --git a/tests/sentry/metrics/test_minimetrics.py b/tests/sentry/metrics/test_minimetrics.py index 8758d816058595..4091b7f4d464ac 100644 --- a/tests/sentry/metrics/test_minimetrics.py +++ b/tests/sentry/metrics/test_minimetrics.py @@ -87,8 +87,8 @@ def test_incr_with_tag(backend, scope): (span,) = scope.client.transport.get_spans() assert span["op"] == "test" - assert span["tags"] == {"x": "y"} assert span["data"]["foo"] == 1 + assert span["data"]["x"] == "y" def test_incr_multi(backend, scope): @@ -101,8 +101,8 @@ def test_incr_multi(backend, scope): (span,) = scope.client.transport.get_spans() assert span["op"] == "test" - assert span["tags"] == {"x": "z"} # NB: Restriction of the interface assert span["data"]["foo"] == 1 # NB: SDK has no get_data() -> incr impossible + assert span["data"]["x"] == "z" def test_gauge(backend, scope): @@ -141,7 +141,7 @@ def test_timing(backend, scope): assert parent["op"] == "test" assert child["op"] == "foo" - assert child["tags"] == {"x": "y"} + assert child["data"]["x"] == "y" duration = datetime.fromisoformat(child["timestamp"]) - datetime.fromisoformat( child["start_timestamp"] @@ -160,8 +160,9 @@ def test_timing_duplicate(backend, scope): (span,) = scope.client.transport.get_spans() assert span["op"] == "test" - assert span["tags"] == {"x": "y"} assert "test" not in span["data"] + assert span["data"]["x"] == "y" + # NB: Explicit timing is discarded
db37d2928a05a94ab10c865a469eb6e6292b924b
2021-03-17 21:59:06
William Mak
fix(trace-view): Fixing truncation when a regex char doesn't exist (#24465)
false
Fixing truncation when a regex char doesn't exist (#24465)
fix
diff --git a/src/sentry/static/sentry/app/components/truncate.tsx b/src/sentry/static/sentry/app/components/truncate.tsx index b4a21d89f8dd23..8d344141ffdef6 100644 --- a/src/sentry/static/sentry/app/components/truncate.tsx +++ b/src/sentry/static/sentry/app/components/truncate.tsx @@ -69,14 +69,14 @@ class Truncate extends React.Component<Props, State> { : value.slice(0, maxLength - 4); // Try to trim to values from the regex - if ( - trimRegex && - leftTrim && - slicedValue.search(trimRegex) <= maxLength - minLength - ) { + if (trimRegex && leftTrim) { + const valueIndex = slicedValue.search(trimRegex); shortValue = ( <span> - … {slicedValue.slice(slicedValue.search(trimRegex), slicedValue.length)} + …{' '} + {valueIndex > 0 && valueIndex <= maxLength - minLength + ? slicedValue.slice(slicedValue.search(trimRegex), slicedValue.length) + : slicedValue} </span> ); } else if (trimRegex && !leftTrim) {
05eb299e6890820e7fecb18af15a6f36e1f93184
2023-01-28 02:39:26
Josh Ferge
fix(replays): catch edge case where project does not exist (#43606)
false
catch edge case where project does not exist (#43606)
fix
diff --git a/src/sentry/replays/usecases/ingest.py b/src/sentry/replays/usecases/ingest.py index b4c1f049ea8970..9ed12e43b528f9 100644 --- a/src/sentry/replays/usecases/ingest.py +++ b/src/sentry/replays/usecases/ingest.py @@ -198,7 +198,18 @@ def ingest_recording(message: RecordingIngestMessage, transaction: Span) -> None # or do this in a separate arroyo step # also need to talk with other teams on only-once produce requirements if headers["segment_id"] == 0 and message.org_id: - project = Project.objects.get_from_cache(id=message.project_id) + try: + project = Project.objects.get_from_cache(id=message.project_id) + except Project.DoesNotExist: + logger.warning( + "Recording segment was received for a project that does not exist.", + extra={ + "project_id": message.project_id, + "replay_id": message.replay_id, + }, + ) + return None + if not project.flags.has_replays: first_replay_received.send_robust(project=project, sender=Project)
01524477886e1cf3eb6b8efe36860ac0eeb007e7
2022-06-07 19:27:22
edwardgou-sentry
feat(releases): updates releases performance card table to use events (#35374)
false
updates releases performance card table to use events (#35374)
feat
diff --git a/static/app/components/discover/performanceCardTable.tsx b/static/app/components/discover/performanceCardTable.tsx index 15eb6ee7d22416..1e38ca184881f7 100644 --- a/static/app/components/discover/performanceCardTable.tsx +++ b/static/app/components/discover/performanceCardTable.tsx @@ -49,7 +49,7 @@ function PerformanceCardTable({ }: PerformanceCardTableProps) { const miseryRenderer = allReleasesTableData?.meta && - getFieldRenderer('user_misery', allReleasesTableData.meta); + getFieldRenderer('user_misery()', allReleasesTableData.meta, false); function renderChange( allReleasesScore: number, @@ -85,8 +85,8 @@ function PerformanceCardTable({ } function userMiseryTrend() { - const allReleasesUserMisery = allReleasesTableData?.data?.[0]?.user_misery; - const thisReleaseUserMisery = thisReleaseTableData?.data?.[0]?.user_misery; + const allReleasesUserMisery = allReleasesTableData?.data?.[0]?.['user_misery()']; + const thisReleaseUserMisery = thisReleaseTableData?.data?.[0]?.['user_misery()']; return ( <StyledPanelItem> {renderChange( @@ -100,16 +100,16 @@ function PerformanceCardTable({ function renderFrontendPerformance() { const webVitals = [ - {title: WebVital.FCP, field: 'p75_measurements_fcp'}, - {title: WebVital.FID, field: 'p75_measurements_fid'}, - {title: WebVital.LCP, field: 'p75_measurements_lcp'}, - {title: WebVital.CLS, field: 'p75_measurements_cls'}, + {title: WebVital.FCP, field: 'p75(measurements.fcp)'}, + {title: WebVital.FID, field: 'p75(measurements.fid)'}, + {title: WebVital.LCP, field: 'p75(measurements.lcp)'}, + {title: WebVital.CLS, field: 'p75(measurements.cls)'}, ]; const spans = [ - {title: 'HTTP', column: 'p75(spans.http)', field: 'p75_spans_http'}, - {title: 'Browser', column: 'p75(spans.browser)', field: 'p75_spans_browser'}, - {title: 'Resource', column: 'p75(spans.resource)', field: 'p75_spans_resource'}, + {title: 'HTTP', column: 'p75(spans.http)', field: 'p75(spans.http)'}, + {title: 'Browser', column: 'p75(spans.browser)', field: 'p75(spans.browser)'}, + {title: 'Resource', column: 'p75(spans.resource)', field: 'p75(spans.resource)'}, ]; const webVitalTitles = webVitals.map((vital, idx) => { @@ -142,13 +142,13 @@ function PerformanceCardTable({ const webVitalsRenderer = webVitals.map( vital => allReleasesTableData?.meta && - getFieldRenderer(vital.field, allReleasesTableData?.meta) + getFieldRenderer(vital.field, allReleasesTableData?.meta, false) ); const spansRenderer = spans.map( span => allReleasesTableData?.meta && - getFieldRenderer(span.field, allReleasesTableData?.meta) + getFieldRenderer(span.field, allReleasesTableData?.meta, false) ); const webReleaseTrend = webVitals.map(vital => { @@ -318,12 +318,13 @@ function PerformanceCardTable({ }); const apdexRenderer = - allReleasesTableData?.meta && getFieldRenderer('apdex', allReleasesTableData.meta); + allReleasesTableData?.meta && + getFieldRenderer('apdex', allReleasesTableData.meta, false); const spansRenderer = spans.map( span => allReleasesTableData?.meta && - getFieldRenderer(span.field, allReleasesTableData?.meta) + getFieldRenderer(span.field, allReleasesTableData?.meta, false) ); const spansReleaseTrend = spans.map(span => { @@ -464,14 +465,15 @@ function PerformanceCardTable({ }); const mobileVitalFields = [ - 'p75_measurements_app_start_cold', - 'p75_measurements_app_start_warm', - 'p75_measurements_frames_slow', - 'p75_measurements_frames_frozen', + 'p75(measurements.app_start_cold)', + 'p75(measurements.app_start_warm)', + 'p75(measurements.frames_slow)', + 'p75(measurements.frames_frozen)', ]; const mobileVitalsRenderer = mobileVitalFields.map( field => - allReleasesTableData?.meta && getFieldRenderer(field, allReleasesTableData?.meta) + allReleasesTableData?.meta && + getFieldRenderer(field, allReleasesTableData?.meta, false) ); const mobileReleaseTrend = mobileVitalFields.map(field => { @@ -699,12 +701,14 @@ function PerformanceCardTableWrapper({ eventView={allReleasesEventView} orgSlug={organization.slug} location={location} + useEvents > {({isLoading, tableData: allReleasesTableData}) => ( <DiscoverQuery eventView={releaseEventView} orgSlug={organization.slug} location={location} + useEvents > {({isLoading: isReleaseLoading, tableData: thisReleaseTableData}) => ( <PerformanceCardTable
7cee40b750f8577242c42de8ea02e65f8fdb91b6
2023-01-04 00:08:58
Colleen O'Rourke
ref(digests): Add generic issues to debug view and tests (#42574)
false
Add generic issues to debug view and tests (#42574)
ref
diff --git a/fixtures/emails/digest.txt b/fixtures/emails/digest.txt index 0e36bf3bddc80a..4c8c754f43dd86 100644 --- a/fixtures/emails/digest.txt +++ b/fixtures/emails/digest.txt @@ -1,21 +1,24 @@ Notifications for example June 22, 2016, 4:16 p.m. UTC to July 4, 2016, 2:18 p.m. UTC -Rule #2 +Rule #3 * DecidingKoiError: faithful hookworm top sharp filly merely sought trivially (8193 events, 9186 users) http://testserver/organizations/example/issues/2/?referrer=digest_email +* something bad happened (5113 events, 5954 users) + http://testserver/organizations/example/issues/201/?referrer=digest_email + * N+1 Query (4418 events, 3851 users) http://testserver/organizations/example/issues/100/?referrer=digest_email -* N+1 Query (2057 events, 2439 users) - http://testserver/organizations/example/issues/101/?referrer=digest_email - * SnailError: molly engaged wahoo feline hedgehog salmon novel boxer polecat (920 events, 5856 users) http://testserver/organizations/example/issues/3/?referrer=digest_email -Rule #3 +* something bad happened (349 events, 6979 users) + http://testserver/organizations/example/issues/200/?referrer=digest_email + +Rule #2 * DecidingKoiError: faithful hookworm top sharp filly merely sought trivially (8193 events, 9186 users) http://testserver/organizations/example/issues/2/?referrer=digest_email @@ -23,20 +26,32 @@ Rule #3 * N+1 Query (4418 events, 3851 users) http://testserver/organizations/example/issues/100/?referrer=digest_email +* N+1 Query (2057 events, 2439 users) + http://testserver/organizations/example/issues/101/?referrer=digest_email + * SnailError: molly engaged wahoo feline hedgehog salmon novel boxer polecat (920 events, 5856 users) http://testserver/organizations/example/issues/3/?referrer=digest_email +* something bad happened (349 events, 6979 users) + http://testserver/organizations/example/issues/200/?referrer=digest_email + Rule #1 * DecidingKoiError: faithful hookworm top sharp filly merely sought trivially (8193 events, 9186 users) http://testserver/organizations/example/issues/2/?referrer=digest_email +* something bad happened (5113 events, 5954 users) + http://testserver/organizations/example/issues/201/?referrer=digest_email + * N+1 Query (4418 events, 3851 users) http://testserver/organizations/example/issues/100/?referrer=digest_email * SnailError: molly engaged wahoo feline hedgehog salmon novel boxer polecat (920 events, 5856 users) http://testserver/organizations/example/issues/3/?referrer=digest_email +* something bad happened (349 events, 6979 users) + http://testserver/organizations/example/issues/200/?referrer=digest_email + Unsubscribe: javascript:alert("This is a preview page, what did you expect to happen?"); diff --git a/src/sentry/web/frontend/debug/debug_generic_issue.py b/src/sentry/web/frontend/debug/debug_generic_issue.py index b0c72270651698..506b900bd061c7 100644 --- a/src/sentry/web/frontend/debug/debug_generic_issue.py +++ b/src/sentry/web/frontend/debug/debug_generic_issue.py @@ -1,49 +1,20 @@ -import uuid -from datetime import datetime - import pytz from django.utils.safestring import mark_safe from django.views.generic import View -from sentry.issues.issue_occurrence import IssueEvidence, IssueOccurrence from sentry.models import Organization, Project, Rule from sentry.notifications.utils import get_generic_data, get_group_settings_link, get_rules -from sentry.types.issues import GroupType from sentry.utils import json -from sentry.utils.dates import ensure_aware -from .mail import COMMIT_EXAMPLE, MailPreview, make_error_event +from .mail import COMMIT_EXAMPLE, MailPreview, make_generic_event class DebugGenericIssueEmailView(View): def get(self, request): - platform = request.GET.get("platform", "python") org = Organization(id=1, slug="example", name="Example") project = Project(id=1, slug="example", name="Example", organization=org) - event = make_error_event(request, project, platform) - event = event.for_group(event.groups[0]) - - occurrence = IssueOccurrence( - uuid.uuid4().hex, - uuid.uuid4().hex, - ["some-fingerprint"], - "something bad happened", - "it was bad", - "1234", - {"Test": 123}, - [ - IssueEvidence("Name 1", "Value 1", True), - IssueEvidence("Name 2", "Value 2", False), - IssueEvidence("Name 3", "Value 3", False), - ], - GroupType.PROFILE_BLOCKED_THREAD, - ensure_aware(datetime.now()), - ) - occurrence.save() - event.occurrence = occurrence - event.group.type = GroupType.PROFILE_BLOCKED_THREAD - + event = make_generic_event(project) group = event.group rule = Rule(id=1, label="An example rule") diff --git a/src/sentry/web/frontend/debug/mail.py b/src/sentry/web/frontend/debug/mail.py index 5a84b004dd722f..44a57fb0f6da06 100644 --- a/src/sentry/web/frontend/debug/mail.py +++ b/src/sentry/web/frontend/debug/mail.py @@ -27,6 +27,7 @@ from sentry.digests.utils import get_digest_metadata from sentry.event_manager import EventManager, get_event_type from sentry.http import get_server_hostname +from sentry.issues.occurrence_consumer import process_event_and_issue_occurrence from sentry.mail.notifications import get_builder_args from sentry.models import ( Activity, @@ -45,6 +46,8 @@ from sentry.notifications.types import GroupSubscriptionReason from sentry.notifications.utils import get_group_settings_link, get_interface_list, get_rules from sentry.testutils.helpers import override_options +from sentry.testutils.helpers.datetime import before_now +from sentry.testutils.helpers.notifications import TEST_ISSUE_OCCURRENCE from sentry.types.issues import GROUP_TYPE_TO_TEXT from sentry.utils import json, loremipsum from sentry.utils.dates import to_datetime, to_timestamp @@ -173,7 +176,6 @@ def make_error_event(request, project, platform): ("level", "error"), ("device", "Other"), ] - event_manager = EventManager(data) event_manager.normalize() data = event_manager.get_data() @@ -208,6 +210,19 @@ def make_performance_event(project): return perf_event +def make_generic_event(project): + occurrence, group_info = process_event_and_issue_occurrence( + TEST_ISSUE_OCCURRENCE.to_dict(), + { + "event_id": uuid.uuid4().hex, + "project_id": project.id, + "timestamp": before_now(minutes=1).isoformat(), + }, + ) + generic_group = group_info.group + return generic_group.get_latest_event() + + def get_shared_context(rule, org, project, group, event): return { "rule": rule, @@ -490,7 +505,6 @@ def digest(request): event = eventstore.create_event( event_id=uuid.uuid4().hex, group_id=group.id, project_id=project.id, data=data.data ) - records.append( Record( event.event_id, @@ -531,6 +545,29 @@ def digest(request): state["event_counts"][perf_group.id] = random.randint(10, 1e4) state["user_counts"][perf_group.id] = random.randint(10, 1e4) + # add in generic issues + for i in range(random.randint(1, 3)): + generic_event = make_generic_event(project) + generic_group = generic_event.group + generic_group.id = i + 200 # don't clobber other issue ids + + records.append( + Record( + generic_event.event_id, + Notification( + generic_event, + random.sample( + list(state["rules"].keys()), random.randint(1, len(state["rules"])) + ), + ), + # this is required for acceptance tests to pass as the EventManager won't accept a timestamp in the past + to_timestamp(datetime(2016, 6, 22, 16, 16, 0, tzinfo=timezone.utc)), + ) + ) + state["groups"][generic_group.id] = generic_group + state["event_counts"][generic_group.id] = random.randint(10, 1e4) + state["user_counts"][generic_group.id] = random.randint(10, 1e4) + digest = build_digest(project, records, state)[0] start, end, counts = get_digest_metadata(digest) diff --git a/tests/sentry/notifications/notifications/test_digests.py b/tests/sentry/notifications/notifications/test_digests.py index 4f32f71c87bf63..017b6ea60ecbc7 100644 --- a/tests/sentry/notifications/notifications/test_digests.py +++ b/tests/sentry/notifications/notifications/test_digests.py @@ -1,3 +1,4 @@ +import uuid from unittest import mock from unittest.mock import patch from urllib.parse import parse_qs @@ -9,40 +10,36 @@ from sentry.digests.backends.base import Backend from sentry.digests.backends.redis import RedisBackend from sentry.digests.notifications import event_to_record -from sentry.event_manager import EventManager +from sentry.issues.occurrence_consumer import process_event_and_issue_occurrence from sentry.models import ProjectOwnership, Rule from sentry.tasks.digests import deliver_digest from sentry.testutils import TestCase -from sentry.testutils.cases import SlackActivityNotificationTest -from sentry.testutils.helpers import override_options +from sentry.testutils.cases import PerformanceIssueTestCase, SlackActivityNotificationTest from sentry.testutils.helpers.datetime import before_now, iso_format from sentry.testutils.helpers.slack import send_notification from sentry.utils import json -from sentry.utils.samples import load_data +from tests.sentry.issues.test_utils import OccurrenceTestMixin USER_COUNT = 2 -class DigestNotificationTest(TestCase): +class DigestNotificationTest(TestCase, OccurrenceTestMixin, PerformanceIssueTestCase): def add_event(self, fingerprint: str, backend: Backend, event_type: str = "error") -> None: if event_type == "performance": - with override_options( + event = self.create_performance_issue() + elif event_type == "generic": + occurrence_data = self.build_occurrence_data() + event_id = uuid.uuid4().hex + occurrence, group_info = process_event_and_issue_occurrence( + occurrence_data, { - "performance.issues.n_plus_one_db.problem-creation": 1.0, - } - ): - data = dict( - load_data( - "transaction-n-plus-one", - timestamp=before_now(days=1), - ) - ) - event_manager = EventManager(data) - event_manager.normalize() - data = event_manager.get_data() - event = event_manager.save(self.project.id) - - event = event.for_group(event.groups[0]) + "event_id": event_id, + "project_id": self.project.id, + "timestamp": before_now(minutes=1).isoformat(), + }, + ) + group = group_info.group + event = group.get_latest_event() else: event = self.store_event( data={ @@ -52,12 +49,19 @@ def add_event(self, fingerprint: str, backend: Backend, event_type: str = "error }, project_id=self.project.id, ) + backend.add( self.key, event_to_record(event, [self.rule]), increment_delay=0, maximum_delay=0 ) @patch.object(sentry, "digests") - def run_test(self, digests, event_count: int, performance_issues: bool = False): + def run_test( + self, + digests, + event_count: int, + performance_issues: bool = False, + generic_issues: bool = False, + ): backend = RedisBackend() digests.digest = backend.digest @@ -65,7 +69,10 @@ def run_test(self, digests, event_count: int, performance_issues: bool = False): self.add_event(f"group-{i}", backend, "error") if performance_issues: - self.add_event(f"group-{i}", backend, "performance") + self.add_event(f"group-{event_count+1}", backend, "performance") + + if generic_issues: + self.add_event(f"group-{event_count+2}", backend, "generic") with self.tasks(): deliver_digest(self.key) @@ -89,10 +96,11 @@ def setUp(self): def test_sends_digest_to_every_member(self): """Test that each member of the project the events are created in receive a digest email notification""" event_count = 4 - self.run_test(event_count=event_count, performance_issues=True) - assert f"{event_count + 1} new alerts since" in mail.outbox[0].subject + self.run_test(event_count=event_count, performance_issues=True, generic_issues=True) + assert f"{event_count + 2} new alerts since" in mail.outbox[0].subject assert "N+1 Query" in mail.outbox[0].body assert "oh no" in mail.outbox[0].body + assert self.build_occurrence_data()["issue_title"] in mail.outbox[0].body def test_sends_alert_rule_notification_to_each_member(self): """Test that if there is only one event it is sent as a regular alert rule notification"""
da6f0a1a56b9ad34f997936bf2539a7e43bc8bda
2021-11-10 04:48:24
Scott Cooper
fix(teamStats): Prevent team misery from showing other teams (#29884)
false
Prevent team misery from showing other teams (#29884)
fix
diff --git a/static/app/views/organizationStats/teamInsights/overview.tsx b/static/app/views/organizationStats/teamInsights/overview.tsx index e0846c37c20d63..dba26cfaaf2923 100644 --- a/static/app/views/organizationStats/teamInsights/overview.tsx +++ b/static/app/views/organizationStats/teamInsights/overview.tsx @@ -262,6 +262,7 @@ function TeamInsightsOverview({location, router}: Props) { <TeamMisery organization={organization} projects={projects} + teamId={currentTeam!.id} period={period} start={start?.toString()} end={end?.toString()} diff --git a/static/app/views/organizationStats/teamInsights/teamMisery.tsx b/static/app/views/organizationStats/teamInsights/teamMisery.tsx index 505454952415ad..8e7fa0575e1161 100644 --- a/static/app/views/organizationStats/teamInsights/teamMisery.tsx +++ b/static/app/views/organizationStats/teamInsights/teamMisery.tsx @@ -166,6 +166,7 @@ function TeamMisery({ type Props = AsyncComponent['props'] & { organization: Organization; + teamId: string; projects: Project[]; location: Location; period?: string; @@ -175,6 +176,7 @@ type Props = AsyncComponent['props'] & { function TeamMiseryWrapper({ organization, + teamId, projects, location, period, @@ -198,9 +200,10 @@ function TeamMiseryWrapper({ const commonEventView = { id: undefined, query: 'transaction.duration:<15m team_key_transaction:true', - projects: projects.map(project => Number(project.id)), + projects: [], version: 2 as SavedQueryVersions, orderby: '-tpm', + teams: [Number(teamId)], fields: [ 'transaction', 'project',
ee2b9283c708672c310b8e08067c735b392d16df
2019-12-18 06:22:00
Danny Lee
feat(discover2): Move buttons to add/edit columns out of the Grid (#15690)
false
Move buttons to add/edit columns out of the Grid (#15690)
feat
diff --git a/src/sentry/static/sentry/app/components/gridEditable/gridHeadCell.tsx b/src/sentry/static/sentry/app/components/gridEditable/gridHeadCell.tsx index 4b2b3f1d4de256..45807afd7548e8 100644 --- a/src/sentry/static/sentry/app/components/gridEditable/gridHeadCell.tsx +++ b/src/sentry/static/sentry/app/components/gridEditable/gridHeadCell.tsx @@ -13,7 +13,6 @@ import { GridHeadCellResizer, } from './styles'; import {GridColumnHeader} from './types'; -import AddColumnButton from './addColumnButton'; export type GridHeadCellProps<Column> = { isColumnDragging: boolean; @@ -132,31 +131,6 @@ class GridHeadCell<Column extends GridColumnHeader> extends React.Component< return <GridHeadCellResizer isEditing={isEditing} />; }; - renderAddColumnButton = () => { - const { - isEditing, - isLast, - openModalAddColumnAt, - indexColumnOrder, - isColumnDragging, - } = this.props; - - if (isLast || !isEditing || isColumnDragging) { - return null; - } - - return ( - <AddColumnButton - align="right" - onClick={() => { - const insertIndex = indexColumnOrder + 1; - openModalAddColumnAt(insertIndex); - }} - data-test-id={`grid-add-column-${indexColumnOrder}`} - /> - ); - }; - render() { const {isEditing, children, column, gridHeadCellButtonProps} = this.props; @@ -179,7 +153,6 @@ class GridHeadCell<Column extends GridColumnHeader> extends React.Component< to ensure that it is will always float on top of everything else */ this.renderResizeGrabbable()} - {this.renderAddColumnButton()} </GridHeadCellWrapper> ); } diff --git a/src/sentry/static/sentry/app/components/gridEditable/index.tsx b/src/sentry/static/sentry/app/components/gridEditable/index.tsx index b706bd4b0545a9..d54149960b34bc 100644 --- a/src/sentry/static/sentry/app/components/gridEditable/index.tsx +++ b/src/sentry/static/sentry/app/components/gridEditable/index.tsx @@ -6,7 +6,6 @@ import {openModal} from 'app/actionCreators/modal'; import EmptyStateWarning from 'app/components/emptyStateWarning'; import InlineSvg from 'app/components/inlineSvg'; import LoadingContainer from 'app/components/loading/loadingContainer'; -import ToolTip from 'app/components/tooltip'; import { GridColumn, @@ -17,10 +16,11 @@ import { } from './types'; import GridHeadCell from './gridHeadCell'; import GridModalEditColumn from './gridModalEditColumn'; -import AddColumnButton from './addColumnButton'; import { - GridPanel, - GridPanelBody, + Header, + HeaderTitle, + HeaderButton, + Body, Grid, GridRow, GridHead, @@ -29,8 +29,6 @@ import { GridBodyCellSpan, GridBodyCellLoading, GridBodyErrorAlert, - GridEditGroup, - GridEditGroupButton, } from './styles'; type GridEditableProps<DataRow, ColumnKey> = { @@ -52,6 +50,7 @@ type GridEditableProps<DataRow, ColumnKey> = { * - `columnSortBy` is not used at the moment, however it might be better to * move sorting into Grid for performance */ + title?: string; columnOrder: GridColumnOrder<ColumnKey>[]; columnSortBy: GridColumnSortBy<ColumnKey>[]; data: DataRow[]; @@ -141,7 +140,14 @@ class GridEditable< this.setState({isEditing: nextValue}); }; - openModalAddColumnAt = (insertIndex: number) => { + /** + * Leave `insertIndex` as undefined to add new column to the end. + */ + openModalAddColumnAt = (insertIndex: number = -1) => { + if (insertIndex < 0) { + insertIndex = this.props.columnOrder.length; + } + return this.toggleModalEditColumn(insertIndex); }; @@ -149,174 +155,109 @@ class GridEditable< indexColumnOrder?: number, column?: GridColumn<ColumnKey> ): void => { - if (this.state.isEditing) { - const {modalEditColumn} = this.props; - - openModal(openModalProps => ( - <GridModalEditColumn - {...openModalProps} - indexColumnOrder={indexColumnOrder} - column={column} - renderBodyWithForm={modalEditColumn.renderBodyWithForm} - renderFooter={modalEditColumn.renderFooter} - /> - )); - } + const {modalEditColumn} = this.props; + + openModal(openModalProps => ( + <GridModalEditColumn + {...openModalProps} + indexColumnOrder={indexColumnOrder} + column={column} + renderBodyWithForm={modalEditColumn.renderBodyWithForm} + renderFooter={modalEditColumn.renderFooter} + /> + )); }; - renderError = () => { - const {error} = this.props; + renderHeaderButton = () => { + if (!this.props.isEditable) { + return null; + } return ( - <React.Fragment> - <GridPanel> - <Grid - isEditable={this.props.isEditable} - isEditing={this.state.isEditing} - numColumn={this.state.numColumn} - > - {this.renderGridHead()} - <GridBody> - <GridRow> - <GridBodyCellSpan> - <GridBodyErrorAlert type="error" icon="icon-circle-exclamation"> - {error} - </GridBodyErrorAlert> - </GridBodyCellSpan> - </GridRow> - </GridBody> - </Grid> - </GridPanel> - </React.Fragment> + <HeaderButton + onClick={() => this.openModalAddColumnAt()} + data-test-id="grid-add-column" + > + <InlineSvg src="icon-circle-add" /> + {t('Add Column')} + </HeaderButton> ); }; - renderLoading = () => { - return ( - <GridBody> - <GridRow> - <GridBodyCellSpan> - <GridBodyCellLoading> - <LoadingContainer isLoading /> - </GridBodyCellLoading> - </GridBodyCellSpan> - </GridRow> - </GridBody> - ); - }; + renderGridHeadEditButtons = () => { + if (!this.props.isEditable) { + return null; + } + + if (!this.state.isEditing) { + return ( + <HeaderButton onClick={this.toggleEdit} data-test-id="grid-edit-enable"> + <InlineSvg src="icon-edit-pencil" /> + {t('Edit Columns')} + </HeaderButton> + ); + } - renderEmptyData = () => { return ( - <GridBody> - <GridRow> - <GridBodyCellSpan> - <EmptyStateWarning> - <p>{t('No results found')}</p> - </EmptyStateWarning> - </GridBodyCellSpan> - </GridRow> - </GridBody> + <HeaderButton onClick={this.toggleEdit} data-test-id="grid-edit-disable"> + <InlineSvg src="icon-circle-close" /> + {t('Exit Edit')} + </HeaderButton> ); }; renderGridHead = () => { - const {isEditable, columnOrder, actions, grid} = this.props; + const {columnOrder, actions, grid} = this.props; const {isEditing} = this.state; // Ensure that the last column cannot be removed const enableEdit = isEditing && columnOrder.length > 1; return ( - <GridHead> - <GridRow> - {/* GridHeadEdit must come first. - - It is a <th> that uses `position: absolute` to set its placement. - The CSS selectors captures the last GridHeadCell and put a - padding-right to provide space for GridHeadEdit to be displayed. - - FAQ: - Instead of using `position: absolute`, why can't we just put - GridHeadEdit at the end so it appears on the right? - Because CSS Grids need to have the same number of Head/Body cells - for everything to align properly. Sub-grids are new and may not be - well supported in older browsers/ - - Why can't we just put GridHeadEdit somewhere else? - Because HTML rules mandate that <div> cannot be a nested child of - a <table>. This seems the best way to make it correct to satisfy - HTML semantics. */ - isEditable && this.renderGridHeadEditButtons()} - - {columnOrder.map((column, columnIndex) => ( - <GridHeadCell - openModalAddColumnAt={this.openModalAddColumnAt} - isLast={columnOrder.length - 1 === columnIndex} - key={`${columnIndex}.${column.key}`} - isColumnDragging={this.props.isColumnDragging} - isPrimary={column.isPrimary} - isEditing={enableEdit} - indexColumnOrder={columnIndex} - column={column} - gridHeadCellButtonProps={this.props.gridHeadCellButtonProps || {}} - actions={{ - moveColumnCommit: actions.moveColumnCommit, - onDragStart: actions.onDragStart, - deleteColumn: actions.deleteColumn, - toggleModalEditColumn: this.toggleModalEditColumn, - }} - > - {grid.renderHeaderCell - ? grid.renderHeaderCell(column, columnIndex) - : column.name} - </GridHeadCell> - ))} - </GridRow> - </GridHead> + <GridRow> + {columnOrder.map((column, columnIndex) => ( + <GridHeadCell + openModalAddColumnAt={this.openModalAddColumnAt} + isLast={columnOrder.length - 1 === columnIndex} + key={`${columnIndex}.${column.key}`} + isColumnDragging={this.props.isColumnDragging} + isPrimary={column.isPrimary} + isEditing={enableEdit} + indexColumnOrder={columnIndex} + column={column} + gridHeadCellButtonProps={this.props.gridHeadCellButtonProps || {}} + actions={{ + moveColumnCommit: actions.moveColumnCommit, + onDragStart: actions.onDragStart, + deleteColumn: actions.deleteColumn, + toggleModalEditColumn: this.toggleModalEditColumn, + }} + > + {grid.renderHeaderCell + ? grid.renderHeaderCell(column, columnIndex) + : column.name} + </GridHeadCell> + ))} + </GridRow> ); }; - renderGridHeadEditButtons = () => { - if (!this.props.isEditable) { - return null; - } + renderGridBody = () => { + const {data, error, isLoading} = this.props; - if (!this.state.isEditing) { - return ( - <GridEditGroup> - <GridEditGroupButton onClick={this.toggleEdit} data-test-id="grid-edit-enable"> - <ToolTip title={t('Edit Columns')}> - <InlineSvg src="icon-edit-pencil" /> - </ToolTip> - </GridEditGroupButton> - </GridEditGroup> - ); + if (error) { + return this.renderError(); } - return ( - <GridEditGroup> - <AddColumnButton - align="left" - onClick={() => this.toggleModalEditColumn()} - data-test-id="grid-add-column-right-end" - /> - <GridEditGroupButton onClick={this.toggleEdit}> - <ToolTip title={t('Exit Edit')}> - <InlineSvg src="icon-close" /> - </ToolTip> - </GridEditGroupButton> - </GridEditGroup> - ); - }; - - renderGridBody = () => { - const {data} = this.props; + if (isLoading) { + return this.renderLoading(); + } if (!data || data.length === 0) { return this.renderEmptyData(); } - return <GridBody>{data.map(this.renderGridBodyRow)}</GridBody>; + return data.map(this.renderGridBodyRow); }; renderGridBodyRow = (dataRow: DataRow, row: number) => { @@ -333,24 +274,76 @@ class GridEditable< ); }; + renderError = () => { + const {error} = this.props; + + return ( + <GridRow> + <GridBodyCellSpan> + <GridBodyErrorAlert type="error" icon="icon-circle-exclamation"> + {error} + </GridBodyErrorAlert> + </GridBodyCellSpan> + </GridRow> + ); + }; + + renderLoading = () => { + return ( + <GridRow> + <GridBodyCellSpan> + <GridBodyCellLoading> + <LoadingContainer isLoading /> + </GridBodyCellLoading> + </GridBodyCellSpan> + </GridRow> + ); + }; + + renderEmptyData = () => { + return ( + <GridRow> + <GridBodyCellSpan> + <EmptyStateWarning> + <p>{t('No results found')}</p> + </EmptyStateWarning> + </GridBodyCellSpan> + </GridRow> + ); + }; + render() { - if (this.props.error) { - return this.renderError(); - } + const {title, isEditable} = this.props; return ( - <GridPanel> - <GridPanelBody> + <React.Fragment> + <Header> + {/* TODO(leedongwei): Check with Bowen/Dora on what they want the + default title to be */} + <HeaderTitle>{title || t('Query Builder')}</HeaderTitle> + + {/* TODO(leedongwei): This is ugly but I need to move it to work on + resizing columns. It will be refactored in a upcoming PR */} + <div style={{display: 'flex', flexDirection: 'row'}}> + {this.renderHeaderButton()} + + <div style={{marginLeft: '16px'}}> + {isEditable && this.renderGridHeadEditButtons()} + </div> + </div> + </Header> + + <Body> <Grid isEditable={this.props.isEditable} isEditing={this.state.isEditing} numColumn={this.state.numColumn} > - {this.renderGridHead()} - {this.props.isLoading ? this.renderLoading() : this.renderGridBody()} + <GridHead>{this.renderGridHead()}</GridHead> + <GridBody>{this.renderGridBody()}</GridBody> </Grid> - </GridPanelBody> - </GridPanel> + </Body> + </React.Fragment> ); } } diff --git a/src/sentry/static/sentry/app/components/gridEditable/styles.tsx b/src/sentry/static/sentry/app/components/gridEditable/styles.tsx index 614e36f980fcd7..baf6999bd34207 100644 --- a/src/sentry/static/sentry/app/components/gridEditable/styles.tsx +++ b/src/sentry/static/sentry/app/components/gridEditable/styles.tsx @@ -1,3 +1,4 @@ +import React from 'react'; import styled from 'react-emotion'; import Alert from 'app/components/alert'; @@ -7,9 +8,6 @@ import space from 'app/styles/space'; export const ADD_BUTTON_SIZE = 16; // this is an even number export const GRID_HEADER_HEIGHT = 45; -const GRID_EDIT_WIDTH = 35; -const GRID_EDIT_WIDTH_EDIT_MODE = - GRID_EDIT_WIDTH + ADD_BUTTON_SIZE / 2 + (12 - ADD_BUTTON_SIZE / 2); /** * Explanation of z-index: @@ -17,8 +15,7 @@ const GRID_EDIT_WIDTH_EDIT_MODE = * - Editable needs to float above Resizer to hide the right-most Resizer, */ const Z_INDEX_RESIZER = '1'; -const Z_INDEX_EDITABLE = '10'; -export const Z_INDEX_ADD_COLUMN = '20'; +export const Z_INDEX_ADD_COLUMN = '20'; // TODO(leedongwei): Remove with addColumnButton.tsx type GridEditableProps = { numColumn?: number; @@ -28,10 +25,42 @@ type GridEditableProps = { isDragging?: boolean; }; -export const GridPanel = styled(Panel)` - /* overflow: hidden; */ +export const Header = styled('div')` + display: flex; + justify-content: space-between; + align-items: center; + + margin-bottom: ${space(1)}; +`; +export const HeaderTitle = styled('h2')` + margin: 0; + + font-size: ${p => p.theme.headerFontSize}; + font-weight: normal; + color: ${p => p.theme.gray3}; `; -export const GridPanelBody = styled(PanelBody)``; +export const HeaderButton = styled('div')` + display: flex; + align-items: center; + + color: ${p => p.theme.gray3}; + cursor: pointer; + + > svg { + margin-right: ${space(1)}; + } + + &:hover, + &:active { + color: ${p => p.theme.gray4}; + } +`; + +export const Body: React.FC = props => ( + <Panel> + <PanelBody>{props.children}</PanelBody> + </Panel> +); /** * @@ -58,25 +87,6 @@ export const Grid = styled('table')<GridEditableProps>` margin: 0; /* background-color: ${p => p.theme.offWhite}; */ - /* overflow: hidden; */ - - /* For the last column, we want to have some space on the right if column - is editable. - - For the header, we set padding for 1 or 2 buttons depending on state - For the body, use "td:last-child" */ - th:last-child { - ${p => { - if (!p.isEditable) { - return 'padding-right: 0px'; - } - if (!p.isEditing) { - return `padding-right: ${GRID_EDIT_WIDTH}px;`; - } - - return `padding-right: ${GRID_EDIT_WIDTH_EDIT_MODE}px;`; - }} - } `; export const GridRow = styled('tr')` display: contents; @@ -315,52 +325,3 @@ export const GridBodyCellLoading = styled('div')` export const GridBodyErrorAlert = styled(Alert)` margin: 0; `; - -/** - * - * GridEditGroup are the buttons that are on the top right of the Grid that - * allows the user to add/remove/resize the columns of the Grid - * - */ -export const GridEditGroup = styled('th')` - position: absolute; - top: 0; - right: 0; - display: flex; - height: ${GRID_HEADER_HEIGHT}px; - - background-color: ${p => p.theme.offWhite}; - border-bottom: 1px solid ${p => p.theme.borderDark}; - border-top-right-radius: ${p => p.theme.borderRadius}; - - z-index: ${Z_INDEX_EDITABLE}; -`; -export const GridEditGroupButton = styled('div')` - display: block; - width: ${GRID_EDIT_WIDTH}px; - height: ${GRID_HEADER_HEIGHT}px; - - color: ${p => p.theme.gray2}; - font-size: 16px; - cursor: pointer; - - &:hover { - color: ${p => p.theme.gray3}; - } - &:active { - color: ${p => p.theme.gray4}; - } - &:last-child { - border-left: 1px solid ${p => p.theme.borderDark}; - } - - /* Targets ToolTip to ensure that it will fill up the parent element and - its child elements will float in its center */ - > span { - display: flex; - justify-content: center; - align-items: center; - width: 100%; - height: 100%; - } -`; diff --git a/src/sentry/static/sentry/app/views/eventsV2/table/index.tsx b/src/sentry/static/sentry/app/views/eventsV2/table/index.tsx index 70bf2cc8ae3eb2..b110d08fde04ca 100644 --- a/src/sentry/static/sentry/app/views/eventsV2/table/index.tsx +++ b/src/sentry/static/sentry/app/views/eventsV2/table/index.tsx @@ -4,6 +4,7 @@ import {browserHistory} from 'react-router'; import styled from 'react-emotion'; import {Client} from 'app/api'; +import space from 'app/styles/space'; import {Organization} from 'app/types'; import withApi from 'app/utils/withApi'; @@ -166,4 +167,5 @@ export default withApi<TableProps>(Table); const Container = styled('div')` min-width: 0; overflow: hidden; + margin-top: ${space(1.5)}; `; diff --git a/tests/acceptance/test_organization_events_v2.py b/tests/acceptance/test_organization_events_v2.py index 94b9758b5c906e..961037e2e948ad 100644 --- a/tests/acceptance/test_organization_events_v2.py +++ b/tests/acceptance/test_organization_events_v2.py @@ -215,7 +215,7 @@ def test_errors_query_empty_state(self): "events-v2 - errors query - empty state - querybuilder - column edit state" ) - self.browser.click_when_visible('[data-test-id="grid-add-column-right-end"]') + self.browser.click_when_visible('[data-test-id="grid-add-column"]') self.browser.snapshot( "events-v2 - errors query - empty state - querybuilder - add column" )
bf51d1d32c3162b24af92ebc2051d1c1bb7ce9a2
2023-06-29 19:23:43
Riccardo Busetti
feat(dynamic-sampling): Improve logging of recalibration (#51867)
false
Improve logging of recalibration (#51867)
feat
diff --git a/src/sentry/dynamic_sampling/tasks/logging.py b/src/sentry/dynamic_sampling/tasks/logging.py index b5114c6f17c88e..378ad8f27d04fc 100644 --- a/src/sentry/dynamic_sampling/tasks/logging.py +++ b/src/sentry/dynamic_sampling/tasks/logging.py @@ -1,5 +1,5 @@ import logging -from typing import Dict, List, Optional +from typing import Optional logger = logging.getLogger(__name__) @@ -41,8 +41,8 @@ def log_query_timeout(query: str, offset: int) -> None: logger.error("dynamic_sampling.query_timeout", extra={"query": query, "offset": offset}) -def log_recalibrate_orgs_errors(errors: Dict[str, List[str]]) -> None: - logger.info("dynamic_sampling.recalibrate_orgs_errors", extra={"errors": errors}) +def log_recalibrate_org_error(org_id: int, error: str) -> None: + logger.info("dynamic_sampling.recalibrate_org_error", extra={"org_id": org_id, "error": error}) def log_recalibrate_org_state( diff --git a/src/sentry/dynamic_sampling/tasks/recalibrate_orgs.py b/src/sentry/dynamic_sampling/tasks/recalibrate_orgs.py index 6cad45d9bbebaa..df4c803e7ef1ac 100644 --- a/src/sentry/dynamic_sampling/tasks/recalibrate_orgs.py +++ b/src/sentry/dynamic_sampling/tasks/recalibrate_orgs.py @@ -1,8 +1,7 @@ import time -from collections import defaultdict from dataclasses import dataclass from datetime import datetime, timedelta -from typing import Dict, Generator, List +from typing import Generator, List from snuba_sdk import ( Column, @@ -31,8 +30,8 @@ set_guarded_adjusted_factor, ) from sentry.dynamic_sampling.tasks.logging import ( + log_recalibrate_org_error, log_recalibrate_org_state, - log_recalibrate_orgs_errors, log_sample_rate_source, ) from sentry.dynamic_sampling.tasks.utils import dynamic_sampling_task @@ -79,17 +78,12 @@ def is_valid_for_recalibration(self): ) @dynamic_sampling_task def recalibrate_orgs() -> None: - errors: Dict[str, List[str]] = defaultdict(list) - for orgs in get_active_orgs(1000): for org_volume in fetch_org_volumes(orgs): try: recalibrate_org(org_volume) except RecalibrationError as e: - errors[str(org_volume.org_id)].append(e.message) - - if errors: - log_recalibrate_orgs_errors(errors=errors) + log_recalibrate_org_error(org_volume.org_id, e.message) def recalibrate_org(org_volume: OrganizationDataVolume) -> None:
2b03fe23f0540638ce94d7a935c116eb2a5f6d54
2021-12-21 00:44:21
Marcos Gaeta
ref(types): Add request types to endpoints/project* (#30751)
false
Add request types to endpoints/project* (#30751)
ref
diff --git a/src/sentry/api/endpoints/project_agnostic_rule_conditions.py b/src/sentry/api/endpoints/project_agnostic_rule_conditions.py index 5899063ac7de44..48e2de907194f2 100644 --- a/src/sentry/api/endpoints/project_agnostic_rule_conditions.py +++ b/src/sentry/api/endpoints/project_agnostic_rule_conditions.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry import features @@ -6,7 +7,7 @@ class ProjectAgnosticRuleConditionsEndpoint(OrganizationEndpoint): - def get(self, request, organization): + def get(self, request: Request, organization) -> Response: """ Retrieve the list of rule conditions """ diff --git a/src/sentry/api/endpoints/project_codeowners_details.py b/src/sentry/api/endpoints/project_codeowners_details.py index e8d4b2dc16126c..826232f4bf9f23 100644 --- a/src/sentry/api/endpoints/project_codeowners_details.py +++ b/src/sentry/api/endpoints/project_codeowners_details.py @@ -2,6 +2,7 @@ from rest_framework import status from rest_framework.exceptions import PermissionDenied +from rest_framework.request import Request from rest_framework.response import Response from sentry import analytics @@ -21,7 +22,7 @@ class ProjectCodeOwnersDetailsEndpoint( ProjectEndpoint, ProjectOwnershipMixin, ProjectCodeOwnersMixin ): def convert_args( - self, request, organization_slug, project_slug, codeowners_id, *args, **kwargs + self, request: Request, organization_slug, project_slug, codeowners_id, *args, **kwargs ): args, kwargs = super().convert_args( request, organization_slug, project_slug, *args, **kwargs @@ -35,7 +36,7 @@ def convert_args( return (args, kwargs) - def put(self, request, project, codeowners): + def put(self, request: Request, project, codeowners) -> Response: """ Update a CodeOwners ````````````` @@ -82,7 +83,7 @@ def put(self, request, project, codeowners): self.track_response_code("update", status.HTTP_400_BAD_REQUEST) return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST) - def delete(self, request, project, codeowners): + def delete(self, request: Request, project, codeowners) -> Response: """ Delete a CodeOwners """ diff --git a/src/sentry/api/endpoints/project_codeowners_request.py b/src/sentry/api/endpoints/project_codeowners_request.py index bead2925582ca3..a63a6603b98bc8 100644 --- a/src/sentry/api/endpoints/project_codeowners_request.py +++ b/src/sentry/api/endpoints/project_codeowners_request.py @@ -1,6 +1,8 @@ import logging from django.utils.translation import ugettext as _ +from rest_framework.request import Request +from rest_framework.response import Response from sentry import roles from sentry.api.bases.project_request_change import ProjectRequestChangeEndpoint @@ -30,7 +32,7 @@ def get_codeowners_request_builder_args(project, recipient, requester_name): class ProjectCodeOwnersRequestEndpoint(ProjectRequestChangeEndpoint): - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Request to Add CODEOWNERS to a Project ```````````````````````````````````` diff --git a/src/sentry/api/endpoints/project_create_sample.py b/src/sentry/api/endpoints/project_create_sample.py index 9e3f111aadcf5c..c3d94447d6a6d3 100644 --- a/src/sentry/api/endpoints/project_create_sample.py +++ b/src/sentry/api/endpoints/project_create_sample.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectEventPermission @@ -11,7 +12,7 @@ class ProjectCreateSampleEndpoint(ProjectEndpoint): # This is the same scope that allows members to view all issues for a project. permission_classes = (ProjectEventPermission,) - def post(self, request, project): + def post(self, request: Request, project) -> Response: event = create_sample_event(project, platform=project.platform, default="javascript") add_group_to_inbox(event.group, GroupInboxReason.NEW) diff --git a/src/sentry/api/endpoints/project_create_sample_transaction.py b/src/sentry/api/endpoints/project_create_sample_transaction.py index a53330db6762d7..f2cfe17343c9d6 100644 --- a/src/sentry/api/endpoints/project_create_sample_transaction.py +++ b/src/sentry/api/endpoints/project_create_sample_transaction.py @@ -3,6 +3,7 @@ from uuid import uuid4 import pytz +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectEventPermission @@ -67,7 +68,7 @@ class ProjectCreateSampleTransactionEndpoint(ProjectEndpoint): # This is the same scope that allows members to view all issues for a project. permission_classes = (ProjectEventPermission,) - def post(self, request, project): + def post(self, request: Request, project) -> Response: samples_root = os.path.join(DATA_ROOT, "samples") with open(os.path.join(samples_root, get_json_name(project))) as fp: data = json.load(fp) diff --git a/src/sentry/api/endpoints/project_details.py b/src/sentry/api/endpoints/project_details.py index f464f526a86743..647cf4b517a15f 100644 --- a/src/sentry/api/endpoints/project_details.py +++ b/src/sentry/api/endpoints/project_details.py @@ -6,6 +6,7 @@ from django.db import IntegrityError, transaction from django.utils import timezone from rest_framework import serializers, status +from rest_framework.request import Request from rest_framework.response import Response from sentry_relay.processing import validate_sampling_condition, validate_sampling_configuration @@ -360,7 +361,7 @@ def _get_unresolved_count(self, project): return queryset.count() - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ Retrieve a Project `````````````````` @@ -385,7 +386,7 @@ def get(self, request, project): return Response(data) - def put(self, request, project): + def put(self, request: Request, project) -> Response: """ Update a Project ```````````````` @@ -752,7 +753,7 @@ def put(self, request, project): return Response(data) @sudo_required - def delete(self, request, project): + def delete(self, request: Request, project) -> Response: """ Delete a Project ```````````````` diff --git a/src/sentry/api/endpoints/project_docs_platform.py b/src/sentry/api/endpoints/project_docs_platform.py index 2d8cee72469e1f..cdecabe1a32850 100644 --- a/src/sentry/api/endpoints/project_docs_platform.py +++ b/src/sentry/api/endpoints/project_docs_platform.py @@ -1,4 +1,5 @@ from django.urls import reverse +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -35,7 +36,7 @@ def replace_keys(html, project_key): class ProjectDocsPlatformEndpoint(ProjectEndpoint): - def get(self, request, project, platform): + def get(self, request: Request, project, platform) -> Response: data = load_doc(platform) if not data: raise ResourceDoesNotExist diff --git a/src/sentry/api/endpoints/project_environment_details.py b/src/sentry/api/endpoints/project_environment_details.py index 9ed3d828f2b95c..32e4d1fbba2e7e 100644 --- a/src/sentry/api/endpoints/project_environment_details.py +++ b/src/sentry/api/endpoints/project_environment_details.py @@ -1,4 +1,5 @@ from rest_framework import serializers +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -12,7 +13,7 @@ class ProjectEnvironmentSerializer(serializers.Serializer): class ProjectEnvironmentDetailsEndpoint(ProjectEndpoint): - def get(self, request, project, environment): + def get(self, request: Request, project, environment) -> Response: try: instance = EnvironmentProject.objects.select_related("environment").get( project=project, @@ -23,7 +24,7 @@ def get(self, request, project, environment): return Response(serialize(instance, request.user)) - def put(self, request, project, environment): + def put(self, request: Request, project, environment) -> Response: try: instance = EnvironmentProject.objects.select_related("environment").get( project=project, diff --git a/src/sentry/api/endpoints/project_environments.py b/src/sentry/api/endpoints/project_environments.py index daee0195f78ddf..2292293444104e 100644 --- a/src/sentry/api/endpoints/project_environments.py +++ b/src/sentry/api/endpoints/project_environments.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -7,7 +8,7 @@ class ProjectEnvironmentsEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's Environments ``````````````````````````````` diff --git a/src/sentry/api/endpoints/project_event_details.py b/src/sentry/api/endpoints/project_event_details.py index 88991d335d78ec..b8f3b0f03a6df0 100644 --- a/src/sentry/api/endpoints/project_event_details.py +++ b/src/sentry/api/endpoints/project_event_details.py @@ -1,6 +1,7 @@ from copy import deepcopy from datetime import datetime +from rest_framework.request import Request from rest_framework.response import Response from sentry import eventstore @@ -9,7 +10,7 @@ class ProjectEventDetailsEndpoint(ProjectEndpoint): - def get(self, request, project, event_id): + def get(self, request: Request, project, event_id) -> Response: """ Retrieve an Event for a Project ``````````````````````````````` @@ -70,8 +71,12 @@ def get(self, request, project, event_id): return Response(data) +from rest_framework.request import Request +from rest_framework.response import Response + + class EventJsonEndpoint(ProjectEndpoint): - def get(self, request, project, event_id): + def get(self, request: Request, project, event_id) -> Response: event = eventstore.get_event_by_id(project.id, event_id) if not event: diff --git a/src/sentry/api/endpoints/project_events.py b/src/sentry/api/endpoints/project_events.py index 46aaa1d513ff77..0252957a87e541 100644 --- a/src/sentry/api/endpoints/project_events.py +++ b/src/sentry/api/endpoints/project_events.py @@ -2,6 +2,8 @@ from functools import partial from django.utils import timezone +from rest_framework.request import Request +from rest_framework.response import Response from sentry import eventstore, features from sentry.api.bases.project import ProjectEndpoint @@ -21,7 +23,7 @@ class ProjectEventsEndpoint(ProjectEndpoint): } @rate_limit_endpoint(limit=5, window=1) - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's Events ``````````````````````` diff --git a/src/sentry/api/endpoints/project_filter_details.py b/src/sentry/api/endpoints/project_filter_details.py index 2920861830ea73..995318cae50990 100644 --- a/src/sentry/api/endpoints/project_filter_details.py +++ b/src/sentry/api/endpoints/project_filter_details.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -7,7 +8,7 @@ class ProjectFilterDetailsEndpoint(ProjectEndpoint): - def put(self, request, project, filter_id): + def put(self, request: Request, project, filter_id) -> Response: """ Update a filter @@ -34,6 +35,7 @@ def put(self, request, project, filter_id): new_state = inbound_filters.set_filter_state(filter_id, project, serializer.validated_data) audit_log_state = AuditLogEntryEvent.PROJECT_ENABLE + returned_state = None if filter_id == "legacy-browsers": if isinstance(current_state, bool) or isinstance(new_state, bool): returned_state = new_state diff --git a/src/sentry/api/endpoints/project_filters.py b/src/sentry/api/endpoints/project_filters.py index 8a390cf3af2ce8..48e84947c4ccd7 100644 --- a/src/sentry/api/endpoints/project_filters.py +++ b/src/sentry/api/endpoints/project_filters.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -5,7 +6,7 @@ class ProjectFiltersEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a project's filters diff --git a/src/sentry/api/endpoints/project_group_index.py b/src/sentry/api/endpoints/project_group_index.py index 229793f770db7e..b9594d8fcc20dd 100644 --- a/src/sentry/api/endpoints/project_group_index.py +++ b/src/sentry/api/endpoints/project_group_index.py @@ -1,5 +1,6 @@ import functools +from rest_framework.request import Request from rest_framework.response import Response from sentry import analytics, eventstore @@ -38,7 +39,7 @@ class ProjectGroupIndexEndpoint(ProjectEndpoint, EnvironmentMixin): @track_slo_response("workflow") @rate_limit_endpoint(limit=3, window=1) - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's Issues ``````````````````````` @@ -170,7 +171,7 @@ def get(self, request, project): return response @track_slo_response("workflow") - def put(self, request, project): + def put(self, request: Request, project) -> Response: """ Bulk Mutate a List of Issues ```````````````````````````` @@ -238,7 +239,7 @@ def put(self, request, project): ) @track_slo_response("workflow") - def delete(self, request, project): + def delete(self, request: Request, project) -> Response: """ Bulk Remove a List of Issues ```````````````````````````` diff --git a/src/sentry/api/endpoints/project_group_stats.py b/src/sentry/api/endpoints/project_group_stats.py index d5d3450279c4d8..b0468eeda8c0c1 100644 --- a/src/sentry/api/endpoints/project_group_stats.py +++ b/src/sentry/api/endpoints/project_group_stats.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.base import EnvironmentMixin, StatsMixin @@ -20,7 +21,7 @@ class ProjectGroupStatsEndpoint(ProjectEndpoint, EnvironmentMixin, StatsMixin): } @rate_limit_endpoint(limit=20, window=1) - def get(self, request, project): + def get(self, request: Request, project) -> Response: try: environment_id = self._get_environment_id_from_request(request, project.organization_id) except Environment.DoesNotExist: diff --git a/src/sentry/api/endpoints/project_grouping_configs.py b/src/sentry/api/endpoints/project_grouping_configs.py index ddf50131ad5dfc..ab0a86b4a7fcf0 100644 --- a/src/sentry/api/endpoints/project_grouping_configs.py +++ b/src/sentry/api/endpoints/project_grouping_configs.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry import features @@ -13,7 +14,7 @@ class ProjectGroupingConfigsEndpoint(ProjectEndpoint): See GroupingConfigsEndpoint """ - def get(self, request, project): + def get(self, request: Request, project) -> Response: configs = [ config.as_dict() for config in sorted(CONFIGURATIONS.values(), key=lambda x: x.id) diff --git a/src/sentry/api/endpoints/project_index.py b/src/sentry/api/endpoints/project_index.py index 7cb84dabed94be..d803e5eecb5d46 100644 --- a/src/sentry/api/endpoints/project_index.py +++ b/src/sentry/api/endpoints/project_index.py @@ -1,6 +1,8 @@ from django.db.models import Q from django.db.models.query import EmptyQuerySet from rest_framework.exceptions import AuthenticationFailed +from rest_framework.request import Request +from rest_framework.response import Response from sentry.api.base import Endpoint from sentry.api.bases.project import ProjectPermission @@ -15,7 +17,7 @@ class ProjectIndexEndpoint(Endpoint): permission_classes = (ProjectPermission,) - def get(self, request): + def get(self, request: Request) -> Response: """ List your Projects `````````````````` diff --git a/src/sentry/api/endpoints/project_issues_resolved_in_release.py b/src/sentry/api/endpoints/project_issues_resolved_in_release.py index e0bb6df4dcf36f..78672fbf5f8943 100644 --- a/src/sentry/api/endpoints/project_issues_resolved_in_release.py +++ b/src/sentry/api/endpoints/project_issues_resolved_in_release.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.base import EnvironmentMixin @@ -11,7 +12,7 @@ class ProjectIssuesResolvedInReleaseEndpoint(ProjectEndpoint, EnvironmentMixin): permission_classes = (ProjectPermission,) - def get(self, request, project, version): + def get(self, request: Request, project, version) -> Response: """ List issues to be resolved in a particular release `````````````````````````````````````````````````` diff --git a/src/sentry/api/endpoints/project_key_details.py b/src/sentry/api/endpoints/project_key_details.py index a8eef3ae170e33..99e921a047261f 100644 --- a/src/sentry/api/endpoints/project_key_details.py +++ b/src/sentry/api/endpoints/project_key_details.py @@ -1,5 +1,6 @@ from django.db.models import F from rest_framework import status +from rest_framework.request import Request from rest_framework.response import Response from sentry import features @@ -12,7 +13,7 @@ class ProjectKeyDetailsEndpoint(ProjectEndpoint): - def get(self, request, project, key_id): + def get(self, request: Request, project, key_id) -> Response: try: key = ProjectKey.objects.get( project=project, public_key=key_id, roles=F("roles").bitor(ProjectKey.roles.store) @@ -22,7 +23,7 @@ def get(self, request, project, key_id): return Response(serialize(key, request.user), status=200) - def put(self, request, project, key_id): + def put(self, request: Request, project, key_id) -> Response: """ Update a Client Key ``````````````````` @@ -90,7 +91,7 @@ def put(self, request, project, key_id): return Response(serialize(key, request.user), status=200) return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST) - def delete(self, request, project, key_id): + def delete(self, request: Request, project, key_id) -> Response: """ Delete a Client Key ``````````````````` diff --git a/src/sentry/api/endpoints/project_key_stats.py b/src/sentry/api/endpoints/project_key_stats.py index b1d5b7c63a1aa5..eb413adceb382d 100644 --- a/src/sentry/api/endpoints/project_key_stats.py +++ b/src/sentry/api/endpoints/project_key_stats.py @@ -1,6 +1,7 @@ from collections import OrderedDict from django.db.models import F +from rest_framework.request import Request from rest_framework.response import Response from sentry import tsdb @@ -23,7 +24,7 @@ class ProjectKeyStatsEndpoint(ProjectEndpoint, StatsMixin): } @rate_limit_endpoint(limit=20, window=1) - def get(self, request, project, key_id): + def get(self, request: Request, project, key_id) -> Response: try: key = ProjectKey.objects.get( project=project, public_key=key_id, roles=F("roles").bitor(ProjectKey.roles.store) diff --git a/src/sentry/api/endpoints/project_keys.py b/src/sentry/api/endpoints/project_keys.py index 20534be3f7f1fa..5f01151f346d01 100644 --- a/src/sentry/api/endpoints/project_keys.py +++ b/src/sentry/api/endpoints/project_keys.py @@ -1,5 +1,6 @@ from django.db.models import F from rest_framework import status +from rest_framework.request import Request from rest_framework.response import Response from sentry import features @@ -10,7 +11,7 @@ class ProjectKeysEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's Client Keys ```````````````````````````` @@ -40,7 +41,7 @@ def get(self, request, project): on_results=lambda x: serialize(x, request.user), ) - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Create a new Client Key ``````````````````````` diff --git a/src/sentry/api/endpoints/project_member_index.py b/src/sentry/api/endpoints/project_member_index.py index fc5e446a3614b4..324ca73b72d91c 100644 --- a/src/sentry/api/endpoints/project_member_index.py +++ b/src/sentry/api/endpoints/project_member_index.py @@ -1,4 +1,5 @@ from django.db.models import Q +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -7,7 +8,7 @@ class ProjectMemberIndexEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: queryset = ( OrganizationMember.objects.filter( Q(user__is_active=True) | Q(user__isnull=True), diff --git a/src/sentry/api/endpoints/project_ownership.py b/src/sentry/api/endpoints/project_ownership.py index 2097a67952bf44..5353723f199b5f 100644 --- a/src/sentry/api/endpoints/project_ownership.py +++ b/src/sentry/api/endpoints/project_ownership.py @@ -1,5 +1,6 @@ from django.utils import timezone from rest_framework import serializers +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -106,7 +107,7 @@ def get_ownership(self, project): class ProjectOwnershipEndpoint(ProjectEndpoint, ProjectOwnershipMixin): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ Retrieve a Project's Ownership configuration ```````````````````````````````````````````` @@ -117,7 +118,7 @@ def get(self, request, project): """ return Response(serialize(self.get_ownership(project), request.user)) - def put(self, request, project): + def put(self, request: Request, project) -> Response: """ Update a Project's Ownership configuration `````````````````````````````````````````` diff --git a/src/sentry/api/endpoints/project_platforms.py b/src/sentry/api/endpoints/project_platforms.py index 89cfacaebba1c4..5a5427035b9f62 100644 --- a/src/sentry/api/endpoints/project_platforms.py +++ b/src/sentry/api/endpoints/project_platforms.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -6,6 +7,6 @@ class ProjectPlatformsEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: queryset = ProjectPlatform.objects.filter(project_id=project.id) return Response(serialize(list(queryset), request.user)) diff --git a/src/sentry/api/endpoints/project_plugin_details.py b/src/sentry/api/endpoints/project_plugin_details.py index fde65d6c536260..34f9287c046010 100644 --- a/src/sentry/api/endpoints/project_plugin_details.py +++ b/src/sentry/api/endpoints/project_plugin_details.py @@ -2,6 +2,7 @@ from django.http.response import Http404 from django.urls import reverse from rest_framework import serializers +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -30,7 +31,7 @@ def _get_plugin(self, plugin_id): except KeyError: raise ResourceDoesNotExist - def get(self, request, project, plugin_id): + def get(self, request: Request, project, plugin_id) -> Response: plugin = self._get_plugin(plugin_id) try: @@ -44,7 +45,7 @@ def get(self, request, project, plugin_id): raise Http404 return Response(context) - def post(self, request, project, plugin_id): + def post(self, request: Request, project, plugin_id) -> Response: """ Enable plugin, Test plugin or Reset plugin values """ @@ -84,7 +85,7 @@ def post(self, request, project, plugin_id): return Response(status=201) - def delete(self, request, project, plugin_id): + def delete(self, request: Request, project, plugin_id) -> Response: """ Disable plugin """ @@ -105,7 +106,7 @@ def delete(self, request, project, plugin_id): return Response(status=204) - def put(self, request, project, plugin_id): + def put(self, request: Request, project, plugin_id) -> Response: plugin = self._get_plugin(plugin_id) config = [ diff --git a/src/sentry/api/endpoints/project_plugins.py b/src/sentry/api/endpoints/project_plugins.py index b9b99a309f436d..7a55471da0d66b 100644 --- a/src/sentry/api/endpoints/project_plugins.py +++ b/src/sentry/api/endpoints/project_plugins.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -7,7 +8,7 @@ class ProjectPluginsEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: context = serialize( [plugin for plugin in plugins.configurable_for_project(project, version=None)], request.user, diff --git a/src/sentry/api/endpoints/project_processingissues.py b/src/sentry/api/endpoints/project_processingissues.py index 77a5e89bd86e1b..0f1a839e566e41 100644 --- a/src/sentry/api/endpoints/project_processingissues.py +++ b/src/sentry/api/endpoints/project_processingissues.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -10,7 +11,7 @@ class ProjectProcessingIssuesDiscardEndpoint(ProjectEndpoint): - def delete(self, request, project): + def delete(self, request: Request, project) -> Response: """ This discards all open processing issues """ @@ -19,7 +20,7 @@ def delete(self, request, project): class ProjectProcessingIssuesFixEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: token = None if request.user_from_signed_request and request.user.is_authenticated: @@ -57,14 +58,14 @@ def get(self, request, project): resp["Content-Type"] = "text/plain" return resp - def permission_denied(self, request, message=None): + def permission_denied(self, request: Request, message=None): resp = render_to_response("sentry/reprocessing-script.sh", {"token": None}) resp["Content-Type"] = "text/plain" return resp class ProjectProcessingIssuesEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a project's processing issues. """ @@ -73,7 +74,7 @@ def get(self, request, project): )[0] return Response(serialize(data, request.user)) - def delete(self, request, project): + def delete(self, request: Request, project) -> Response: """ This deletes all open processing issues and triggers reprocessing if the user disabled the checkbox diff --git a/src/sentry/api/endpoints/project_release_commits.py b/src/sentry/api/endpoints/project_release_commits.py index 0a62e5ce7f9efd..d0520988d6f901 100644 --- a/src/sentry/api/endpoints/project_release_commits.py +++ b/src/sentry/api/endpoints/project_release_commits.py @@ -1,3 +1,6 @@ +from rest_framework.request import Request +from rest_framework.response import Response + from sentry.api.bases.project import ProjectEndpoint, ProjectReleasePermission from sentry.api.exceptions import ResourceDoesNotExist from sentry.api.serializers import serialize @@ -7,7 +10,7 @@ class ProjectReleaseCommitsEndpoint(ProjectEndpoint): permission_classes = (ProjectReleasePermission,) - def get(self, request, project, version): + def get(self, request: Request, project, version) -> Response: """ List a Project Release's Commits ```````````````````````````````` diff --git a/src/sentry/api/endpoints/project_release_details.py b/src/sentry/api/endpoints/project_release_details.py index 8664568178d86a..4d43a541887f3b 100644 --- a/src/sentry/api/endpoints/project_release_details.py +++ b/src/sentry/api/endpoints/project_release_details.py @@ -1,4 +1,5 @@ from rest_framework.exceptions import ParseError +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.base import ReleaseAnalyticsMixin @@ -17,7 +18,7 @@ class ProjectReleaseDetailsEndpoint(ProjectEndpoint, ReleaseAnalyticsMixin): permission_classes = (ProjectReleasePermission,) - def get(self, request, project, version): + def get(self, request: Request, project, version) -> Response: """ Retrieve a Project's Release ```````````````````````````` @@ -60,7 +61,7 @@ def get(self, request, project, version): ) ) - def put(self, request, project, version): + def put(self, request: Request, project, version) -> Response: """ Update a Project's Release `````````````````````````` @@ -137,7 +138,7 @@ def put(self, request, project, version): return Response(serialize(release, request.user)) - def delete(self, request, project, version): + def delete(self, request: Request, project, version) -> Response: """ Delete a Project's Release `````````````````````````` diff --git a/src/sentry/api/endpoints/project_release_file_details.py b/src/sentry/api/endpoints/project_release_file_details.py index d0ede5558b5b4d..e7aa29cbbb766c 100644 --- a/src/sentry/api/endpoints/project_release_file_details.py +++ b/src/sentry/api/endpoints/project_release_file_details.py @@ -5,6 +5,7 @@ from django.http.response import FileResponse from rest_framework import serializers from rest_framework.exceptions import ParseError +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectReleasePermission @@ -177,7 +178,7 @@ def delete_releasefile(cls, release, file_id): class ProjectReleaseFileDetailsEndpoint(ProjectEndpoint, ReleaseFileDetailsMixin): permission_classes = (ProjectReleasePermission,) - def get(self, request, project, version, file_id): + def get(self, request: Request, project, version, file_id) -> Response: """ Retrieve a Project Release's File ````````````````````````````````` @@ -208,7 +209,7 @@ def get(self, request, project, version, file_id): check_permission_fn=lambda: has_download_permission(request, project), ) - def put(self, request, project, version, file_id): + def put(self, request: Request, project, version, file_id) -> Response: """ Update a File ````````````` @@ -235,7 +236,7 @@ def put(self, request, project, version, file_id): return self.update_releasefile(request, release, file_id) - def delete(self, request, project, version, file_id): + def delete(self, request: Request, project, version, file_id) -> Response: """ Delete a File ````````````` diff --git a/src/sentry/api/endpoints/project_release_files.py b/src/sentry/api/endpoints/project_release_files.py index 19bdfc0c711f24..ee5564100874c8 100644 --- a/src/sentry/api/endpoints/project_release_files.py +++ b/src/sentry/api/endpoints/project_release_files.py @@ -5,6 +5,7 @@ from django.db import IntegrityError, router from django.db.models import Q from django.utils.functional import cached_property +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectReleasePermission @@ -44,7 +45,7 @@ def load_dist(results): class ReleaseFilesMixin: - def get_releasefiles(self, request, release, organization_id): + def get_releasefiles(self, request: Request, release, organization_id): query = request.GET.getlist("query") data_sources = [] @@ -208,7 +209,7 @@ def pseudo_releasefile(url, info, dist): class ProjectReleaseFilesEndpoint(ProjectEndpoint, ReleaseFilesMixin): permission_classes = (ProjectReleasePermission,) - def get(self, request, project, version): + def get(self, request: Request, project, version) -> Response: """ List a Project Release's Files `````````````````````````````` @@ -232,7 +233,7 @@ def get(self, request, project, version): return self.get_releasefiles(request, release, project.organization_id) - def post(self, request, project, version): + def post(self, request: Request, project, version) -> Response: """ Upload a New Project Release File ````````````````````````````````` diff --git a/src/sentry/api/endpoints/project_release_repositories.py b/src/sentry/api/endpoints/project_release_repositories.py index 48169c9f0d6222..b89eb4cabd2bb1 100644 --- a/src/sentry/api/endpoints/project_release_repositories.py +++ b/src/sentry/api/endpoints/project_release_repositories.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectReleasePermission @@ -10,7 +11,7 @@ class ProjectReleaseRepositories(ProjectEndpoint): permission_classes = (ProjectReleasePermission,) - def get(self, request, project, version): + def get(self, request: Request, project, version) -> Response: """ Retrieve Project Repositories from a Release ```````````````````````````` diff --git a/src/sentry/api/endpoints/project_release_setup.py b/src/sentry/api/endpoints/project_release_setup.py index 0cc76cc3a1d7d4..dfe08d0ea00a36 100644 --- a/src/sentry/api/endpoints/project_release_setup.py +++ b/src/sentry/api/endpoints/project_release_setup.py @@ -1,4 +1,5 @@ from django.core.cache import cache +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectReleasePermission @@ -9,7 +10,7 @@ class ProjectReleaseSetupCompletionEndpoint(ProjectEndpoint): permission_classes = (ProjectReleasePermission,) - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ Get list with release setup progress for a project 1. tag an error diff --git a/src/sentry/api/endpoints/project_release_stats.py b/src/sentry/api/endpoints/project_release_stats.py index 55c0158ba33f29..82d6f51feedd04 100644 --- a/src/sentry/api/endpoints/project_release_stats.py +++ b/src/sentry/api/endpoints/project_release_stats.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry import release_health @@ -24,7 +25,7 @@ def upsert_missing_release(project, version): class ProjectReleaseStatsEndpoint(ProjectEndpoint): permission_classes = (ProjectReleasePermission,) - def get(self, request, project, version): + def get(self, request: Request, project, version) -> Response: """ Get a Project Release's Stats ````````````````````````````` diff --git a/src/sentry/api/endpoints/project_releases.py b/src/sentry/api/endpoints/project_releases.py index cfdb53b26b068c..25c5f8e8ca29b4 100644 --- a/src/sentry/api/endpoints/project_releases.py +++ b/src/sentry/api/endpoints/project_releases.py @@ -1,5 +1,6 @@ from django.db import IntegrityError, transaction from django.db.models import Q +from rest_framework.request import Request from rest_framework.response import Response from sentry import analytics @@ -17,7 +18,7 @@ class ProjectReleasesEndpoint(ProjectEndpoint, EnvironmentMixin): permission_classes = (ProjectReleasePermission,) - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's Releases ````````````````````````` @@ -67,7 +68,7 @@ def get(self, request, project): ), ) - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Create a New Release for a Project `````````````````````````````````` diff --git a/src/sentry/api/endpoints/project_releases_token.py b/src/sentry/api/endpoints/project_releases_token.py index 0f916c4db4de32..230231d9b79ce7 100644 --- a/src/sentry/api/endpoints/project_releases_token.py +++ b/src/sentry/api/endpoints/project_releases_token.py @@ -3,6 +3,7 @@ from uuid import uuid1 from django.urls import reverse +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, StrictProjectPermission @@ -40,7 +41,7 @@ def _regenerate_token(self, project): ProjectOption.objects.set_value(project, "sentry:release-token", token) return token - def get(self, request, project): + def get(self, request: Request, project) -> Response: token = ProjectOption.objects.get_value(project, "sentry:release-token") if token is None: @@ -48,7 +49,7 @@ def get(self, request, project): return Response({"token": token, "webhookUrl": _get_webhook_url(project, "builtin", token)}) - def post(self, request, project): + def post(self, request: Request, project) -> Response: token = self._regenerate_token(project) return Response({"token": token, "webhookUrl": _get_webhook_url(project, "builtin", token)}) diff --git a/src/sentry/api/endpoints/project_repo_path_parsing.py b/src/sentry/api/endpoints/project_repo_path_parsing.py index 8e82e12877655b..6e122e18efb187 100644 --- a/src/sentry/api/endpoints/project_repo_path_parsing.py +++ b/src/sentry/api/endpoints/project_repo_path_parsing.py @@ -1,4 +1,6 @@ from rest_framework import serializers, status +from rest_framework.request import Request +from rest_framework.response import Response from sentry import integrations from sentry.api.bases.project import ProjectEndpoint @@ -97,7 +99,7 @@ class ProjectRepoPathParsingEndpoint(ProjectEndpoint): depending on the source code URL """ - def post(self, request, project): + def post(self, request: Request, project) -> Response: serializer = PathMappingSerializer( context={"organization_id": project.organization_id}, data=request.data, diff --git a/src/sentry/api/endpoints/project_reprocessing.py b/src/sentry/api/endpoints/project_reprocessing.py index 5357868661541f..becc1b0a57bcc2 100644 --- a/src/sentry/api/endpoints/project_reprocessing.py +++ b/src/sentry/api/endpoints/project_reprocessing.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectReleasePermission @@ -7,7 +8,7 @@ class ProjectReprocessingEndpoint(ProjectEndpoint): permission_classes = (ProjectReleasePermission,) - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Triggers the reprocessing process as a task """ diff --git a/src/sentry/api/endpoints/project_rule_details.py b/src/sentry/api/endpoints/project_rule_details.py index 89c7e7287d7890..83ed87eb35fcbb 100644 --- a/src/sentry/api/endpoints/project_rule_details.py +++ b/src/sentry/api/endpoints/project_rule_details.py @@ -1,4 +1,5 @@ from rest_framework import status +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectAlertRulePermission, ProjectEndpoint @@ -23,7 +24,7 @@ class ProjectRuleDetailsEndpoint(ProjectEndpoint): permission_classes = (ProjectAlertRulePermission,) - def convert_args(self, request, rule_id, *args, **kwargs): + def convert_args(self, request: Request, rule_id, *args, **kwargs): args, kwargs = super().convert_args(request, *args, **kwargs) project = kwargs["project"] @@ -40,7 +41,7 @@ def convert_args(self, request, rule_id, *args, **kwargs): return args, kwargs @transaction_start("ProjectRuleDetailsEndpoint") - def get(self, request, project, rule): + def get(self, request: Request, project, rule) -> Response: """ Retrieve a rule @@ -54,7 +55,7 @@ def get(self, request, project, rule): return Response(data) @transaction_start("ProjectRuleDetailsEndpoint") - def put(self, request, project, rule): + def put(self, request: Request, project, rule) -> Response: """ Update a rule @@ -133,7 +134,7 @@ def put(self, request, project, rule): return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST) @transaction_start("ProjectRuleDetailsEndpoint") - def delete(self, request, project, rule): + def delete(self, request: Request, project, rule) -> Response: """ Delete a rule """ diff --git a/src/sentry/api/endpoints/project_rule_task_details.py b/src/sentry/api/endpoints/project_rule_task_details.py index 93ad4326fbdaab..a7754806b2d732 100644 --- a/src/sentry/api/endpoints/project_rule_task_details.py +++ b/src/sentry/api/endpoints/project_rule_task_details.py @@ -1,4 +1,5 @@ from django.http import Http404 +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectSettingPermission @@ -10,7 +11,7 @@ class ProjectRuleTaskDetailsEndpoint(ProjectEndpoint): permission_classes = [ProjectSettingPermission] - def get(self, request, project, task_uuid): + def get(self, request: Request, project, task_uuid) -> Response: """ Retrieve the status of the async task diff --git a/src/sentry/api/endpoints/project_rules.py b/src/sentry/api/endpoints/project_rules.py index e0173da613984c..fdb2c4e227619a 100644 --- a/src/sentry/api/endpoints/project_rules.py +++ b/src/sentry/api/endpoints/project_rules.py @@ -1,6 +1,7 @@ from typing import Mapping, Optional, Sequence from rest_framework import serializers, status +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectAlertRulePermission, ProjectEndpoint @@ -49,7 +50,7 @@ class ProjectRulesEndpoint(ProjectEndpoint): permission_classes = (ProjectAlertRulePermission,) @transaction_start("ProjectRulesEndpoint") - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a project's rules @@ -70,7 +71,7 @@ def get(self, request, project): ) @transaction_start("ProjectRulesEndpoint") - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Create a rule diff --git a/src/sentry/api/endpoints/project_rules_configuration.py b/src/sentry/api/endpoints/project_rules_configuration.py index d11da72920e525..86a11202befa1b 100644 --- a/src/sentry/api/endpoints/project_rules_configuration.py +++ b/src/sentry/api/endpoints/project_rules_configuration.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry import features @@ -7,7 +8,7 @@ class ProjectRulesConfigurationEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ Retrieve the list of configuration options for a given project. """ diff --git a/src/sentry/api/endpoints/project_search_details.py b/src/sentry/api/endpoints/project_search_details.py index 400d60f4beda1a..245d65348e92fb 100644 --- a/src/sentry/api/endpoints/project_search_details.py +++ b/src/sentry/api/endpoints/project_search_details.py @@ -1,4 +1,5 @@ from rest_framework import serializers +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, RelaxedSearchPermission @@ -21,7 +22,7 @@ class SavedSearchSerializer(serializers.Serializer): class ProjectSearchDetailsEndpoint(ProjectEndpoint): permission_classes = (RelaxedSearchPermission,) - def get(self, request, project, search_id): + def get(self, request: Request, project, search_id) -> Response: """ Retrieve a saved search @@ -37,7 +38,7 @@ def get(self, request, project, search_id): return Response(serialize(search, request.user)) - def put(self, request, project, search_id): + def put(self, request: Request, project, search_id) -> Response: """ Update a saved search @@ -92,7 +93,7 @@ def put(self, request, project, search_id): return Response(serialize(search, request.user)) - def delete(self, request, project, search_id): + def delete(self, request: Request, project, search_id) -> Response: """ Delete a saved search diff --git a/src/sentry/api/endpoints/project_searches.py b/src/sentry/api/endpoints/project_searches.py index ec4c3004c7e454..405ded7d965ed6 100644 --- a/src/sentry/api/endpoints/project_searches.py +++ b/src/sentry/api/endpoints/project_searches.py @@ -1,6 +1,7 @@ from django.db import IntegrityError, transaction from django.db.models import Q from rest_framework import serializers +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, RelaxedSearchPermission @@ -19,7 +20,7 @@ class SavedSearchSerializer(serializers.Serializer): class ProjectSearchesEndpoint(ProjectEndpoint): permission_classes = (RelaxedSearchPermission,) - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a project's saved searches @@ -36,7 +37,7 @@ def get(self, request, project): return Response(serialize(results, request.user)) - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Create a new saved search diff --git a/src/sentry/api/endpoints/project_servicehook_details.py b/src/sentry/api/endpoints/project_servicehook_details.py index 8a87be0bfc887e..b2d48813314bca 100644 --- a/src/sentry/api/endpoints/project_servicehook_details.py +++ b/src/sentry/api/endpoints/project_servicehook_details.py @@ -1,5 +1,7 @@ from django.db import transaction from rest_framework import status +from rest_framework.request import Request +from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint from sentry.api.exceptions import ResourceDoesNotExist @@ -10,7 +12,7 @@ class ProjectServiceHookDetailsEndpoint(ProjectEndpoint): - def get(self, request, project, hook_id): + def get(self, request: Request, project, hook_id) -> Response: """ Retrieve a Service Hook ``````````````````````` @@ -30,7 +32,7 @@ def get(self, request, project, hook_id): raise ResourceDoesNotExist return self.respond(serialize(hook, request.user)) - def put(self, request, project, hook_id): + def put(self, request: Request, project, hook_id) -> Response: """ Update a Service Hook ````````````````````` @@ -83,7 +85,7 @@ def put(self, request, project, hook_id): return self.respond(serialize(hook, request.user)) - def delete(self, request, project, hook_id): + def delete(self, request: Request, project, hook_id) -> Response: """ Remove a Service Hook ````````````````````` diff --git a/src/sentry/api/endpoints/project_servicehook_stats.py b/src/sentry/api/endpoints/project_servicehook_stats.py index d7de2790cbf110..7f6b4afd87bec7 100644 --- a/src/sentry/api/endpoints/project_servicehook_stats.py +++ b/src/sentry/api/endpoints/project_servicehook_stats.py @@ -1,5 +1,8 @@ from collections import OrderedDict +from rest_framework.request import Request +from rest_framework.response import Response + from sentry import tsdb from sentry.api.base import StatsMixin from sentry.api.bases.project import ProjectEndpoint @@ -8,7 +11,7 @@ class ProjectServiceHookStatsEndpoint(ProjectEndpoint, StatsMixin): - def get(self, request, project, hook_id): + def get(self, request: Request, project, hook_id) -> Response: try: hook = ServiceHook.objects.get(project_id=project.id, guid=hook_id) except ServiceHook.DoesNotExist: diff --git a/src/sentry/api/endpoints/project_servicehooks.py b/src/sentry/api/endpoints/project_servicehooks.py index ca8af2778c393e..cc46d7b5bc712a 100644 --- a/src/sentry/api/endpoints/project_servicehooks.py +++ b/src/sentry/api/endpoints/project_servicehooks.py @@ -1,5 +1,7 @@ from django.db import transaction from rest_framework import status +from rest_framework.request import Request +from rest_framework.response import Response from sentry import features from sentry.api.bases.project import ProjectEndpoint @@ -10,10 +12,10 @@ class ProjectServiceHooksEndpoint(ProjectEndpoint): - def has_feature(self, request, project): + def has_feature(self, request: Request, project): return features.has("projects:servicehooks", project=project, actor=request.user) - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's Service Hooks `````````````````````````````` @@ -54,7 +56,7 @@ def get(self, request, project): on_results=lambda x: serialize(x, request.user), ) - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Register a new Service Hook ``````````````````````````` diff --git a/src/sentry/api/endpoints/project_stacktrace_link.py b/src/sentry/api/endpoints/project_stacktrace_link.py index 9317c93e1d22e9..52f061a6a0fd7b 100644 --- a/src/sentry/api/endpoints/project_stacktrace_link.py +++ b/src/sentry/api/endpoints/project_stacktrace_link.py @@ -1,5 +1,6 @@ from typing import Optional, Tuple +from rest_framework.request import Request from rest_framework.response import Response from sentry_sdk import configure_scope @@ -47,7 +48,7 @@ class ProjectStacktraceLinkEndpoint(ProjectEndpoint): """ - def get(self, request, project): + def get(self, request: Request, project) -> Response: # should probably feature gate filepath = request.GET.get("file") if not filepath: diff --git a/src/sentry/api/endpoints/project_stats.py b/src/sentry/api/endpoints/project_stats.py index 38a1a73f55708a..c8cca493a49d8e 100644 --- a/src/sentry/api/endpoints/project_stats.py +++ b/src/sentry/api/endpoints/project_stats.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry import tsdb @@ -9,7 +10,7 @@ class ProjectStatsEndpoint(ProjectEndpoint, EnvironmentMixin, StatsMixin): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ Retrieve Event Counts for a Project ``````````````````````````````````` diff --git a/src/sentry/api/endpoints/project_tagkey_details.py b/src/sentry/api/endpoints/project_tagkey_details.py index b809d675939819..2fe699c99d9fe1 100644 --- a/src/sentry/api/endpoints/project_tagkey_details.py +++ b/src/sentry/api/endpoints/project_tagkey_details.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry import tagstore @@ -10,7 +11,7 @@ class ProjectTagKeyDetailsEndpoint(ProjectEndpoint, EnvironmentMixin): - def get(self, request, project, key): + def get(self, request: Request, project, key) -> Response: lookup_key = tagstore.prefix_reserved_key(key) try: @@ -26,7 +27,7 @@ def get(self, request, project, key): return Response(serialize(tagkey, request.user)) - def delete(self, request, project, key): + def delete(self, request: Request, project, key) -> Response: """ Remove all occurrences of the given tag key. diff --git a/src/sentry/api/endpoints/project_tagkey_values.py b/src/sentry/api/endpoints/project_tagkey_values.py index 267b2093c83668..5317b159ddf581 100644 --- a/src/sentry/api/endpoints/project_tagkey_values.py +++ b/src/sentry/api/endpoints/project_tagkey_values.py @@ -1,3 +1,6 @@ +from rest_framework.request import Request +from rest_framework.response import Response + from sentry import tagstore from sentry.api.base import EnvironmentMixin from sentry.api.bases.project import ProjectEndpoint @@ -8,7 +11,7 @@ class ProjectTagKeyValuesEndpoint(ProjectEndpoint, EnvironmentMixin): - def get(self, request, project, key): + def get(self, request: Request, project, key) -> Response: """ List a Tag's Values ``````````````````` diff --git a/src/sentry/api/endpoints/project_tags.py b/src/sentry/api/endpoints/project_tags.py index 3279b7b22e433d..124cd5d7a16eda 100644 --- a/src/sentry/api/endpoints/project_tags.py +++ b/src/sentry/api/endpoints/project_tags.py @@ -1,3 +1,4 @@ +from rest_framework.request import Request from rest_framework.response import Response from sentry import tagstore @@ -8,7 +9,7 @@ class ProjectTagsEndpoint(ProjectEndpoint, EnvironmentMixin): - def get(self, request, project): + def get(self, request: Request, project) -> Response: try: environment_id = self._get_environment_id_from_request(request, project.organization_id) except Environment.DoesNotExist: diff --git a/src/sentry/api/endpoints/project_team_details.py b/src/sentry/api/endpoints/project_team_details.py index ff810e958a1cf7..82fe6f4be4ab21 100644 --- a/src/sentry/api/endpoints/project_team_details.py +++ b/src/sentry/api/endpoints/project_team_details.py @@ -1,4 +1,5 @@ from django.http import Http404 +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint, ProjectPermission @@ -21,7 +22,7 @@ class ProjectTeamsPermission(ProjectPermission): class ProjectTeamDetailsEndpoint(ProjectEndpoint): permission_classes = (ProjectTeamsPermission,) - def post(self, request, project, team_slug): + def post(self, request: Request, project, team_slug) -> Response: """ Give a team access to a project ``````````````````````````````` @@ -41,7 +42,7 @@ def post(self, request, project, team_slug): project.add_team(team) return Response(serialize(project, request.user, ProjectWithTeamSerializer()), status=201) - def delete(self, request, project, team_slug): + def delete(self, request: Request, project, team_slug) -> Response: """ Revoke a team's access to a project ``````````````````````````````````` diff --git a/src/sentry/api/endpoints/project_teams.py b/src/sentry/api/endpoints/project_teams.py index 9cfd2021916263..3d1b1b3f0c69c8 100644 --- a/src/sentry/api/endpoints/project_teams.py +++ b/src/sentry/api/endpoints/project_teams.py @@ -1,3 +1,6 @@ +from rest_framework.request import Request +from rest_framework.response import Response + from sentry.api.bases.project import ProjectEndpoint from sentry.api.paginator import OffsetPaginator from sentry.api.serializers import serialize @@ -5,7 +8,7 @@ class ProjectTeamsEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's Teams `````````````````````` diff --git a/src/sentry/api/endpoints/project_transaction_threshold.py b/src/sentry/api/endpoints/project_transaction_threshold.py index aba4f2ee5a61a6..f07cef7160b5b6 100644 --- a/src/sentry/api/endpoints/project_transaction_threshold.py +++ b/src/sentry/api/endpoints/project_transaction_threshold.py @@ -1,5 +1,6 @@ from django.db import transaction from rest_framework import serializers, status +from rest_framework.request import Request from rest_framework.response import Response from sentry import features @@ -45,7 +46,7 @@ def has_feature(self, project, request): "organizations:performance-view", project.organization, actor=request.user ) - def get(self, request, project): + def get(self, request: Request, project) -> Response: if not self.has_feature(project, request): return self.respond(status=status.HTTP_404_NOT_FOUND) @@ -68,7 +69,7 @@ def get(self, request, project): status.HTTP_200_OK, ) - def post(self, request, project): + def post(self, request: Request, project) -> Response: if not self.has_feature(project, request): return self.respond(status=status.HTTP_404_NOT_FOUND) @@ -111,7 +112,7 @@ def post(self, request, project): status=status.HTTP_201_CREATED if created else status.HTTP_200_OK, ) - def delete(self, request, project): + def delete(self, request: Request, project) -> Response: if not self.has_feature(project, request): return self.respond(status=status.HTTP_404_NOT_FOUND) diff --git a/src/sentry/api/endpoints/project_transaction_threshold_override.py b/src/sentry/api/endpoints/project_transaction_threshold_override.py index 381765219c9449..81ecc5bbf9e928 100644 --- a/src/sentry/api/endpoints/project_transaction_threshold_override.py +++ b/src/sentry/api/endpoints/project_transaction_threshold_override.py @@ -1,6 +1,7 @@ from django.db import transaction from rest_framework import serializers, status from rest_framework.exceptions import ParseError +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases import ProjectTransactionThresholdOverridePermission @@ -55,14 +56,14 @@ def validate(self, data): class ProjectTransactionThresholdOverrideEndpoint(OrganizationEventsV2EndpointBase): permission_classes = (ProjectTransactionThresholdOverridePermission,) - def get_project(self, request, organization): + def get_project(self, request: Request, organization): projects = self.get_projects(request, organization) if len(projects) != 1: raise ParseError("Only 1 project per transaction threshold") return projects[0] - def get(self, request, organization): + def get(self, request: Request, organization) -> Response: if not self.has_feature(organization, request): return self.respond(status=status.HTTP_404_NOT_FOUND) @@ -85,7 +86,7 @@ def get(self, request, organization): status.HTTP_200_OK, ) - def post(self, request, organization): + def post(self, request: Request, organization) -> Response: if not self.has_feature(organization, request): return self.respond(status=status.HTTP_404_NOT_FOUND) @@ -126,7 +127,7 @@ def post(self, request, organization): status=status.HTTP_201_CREATED if created else status.HTTP_200_OK, ) - def delete(self, request, organization): + def delete(self, request: Request, organization) -> Response: if not self.has_feature(organization, request): return self.respond(status=status.HTTP_404_NOT_FOUND) diff --git a/src/sentry/api/endpoints/project_transfer.py b/src/sentry/api/endpoints/project_transfer.py index d544716ce58688..82bb6451d509db 100644 --- a/src/sentry/api/endpoints/project_transfer.py +++ b/src/sentry/api/endpoints/project_transfer.py @@ -4,6 +4,7 @@ from django.utils import timezone from rest_framework import status +from rest_framework.request import Request from rest_framework.response import Response from sentry import options, roles @@ -25,7 +26,7 @@ class ProjectTransferEndpoint(ProjectEndpoint): permission_classes = [RelaxedProjectPermission] @sudo_required - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Transfer a Project ```````````````` diff --git a/src/sentry/api/endpoints/project_user_details.py b/src/sentry/api/endpoints/project_user_details.py index 02293a6f582f61..d3e03a766a879d 100644 --- a/src/sentry/api/endpoints/project_user_details.py +++ b/src/sentry/api/endpoints/project_user_details.py @@ -1,4 +1,5 @@ from rest_framework import status +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -8,11 +9,11 @@ class ProjectUserDetailsEndpoint(ProjectEndpoint): - def get(self, request, project, user_hash): + def get(self, request: Request, project, user_hash) -> Response: euser = EventUser.objects.get(project_id=project.id, hash=user_hash) return Response(serialize(euser, request.user)) - def delete(self, request, project, user_hash): + def delete(self, request: Request, project, user_hash) -> Response: """ Delete an Event User ```````````````````````````````` diff --git a/src/sentry/api/endpoints/project_user_reports.py b/src/sentry/api/endpoints/project_user_reports.py index 8bac4714a92412..12dc414321f6f1 100644 --- a/src/sentry/api/endpoints/project_user_reports.py +++ b/src/sentry/api/endpoints/project_user_reports.py @@ -1,4 +1,6 @@ from rest_framework import serializers +from rest_framework.request import Request +from rest_framework.response import Response from sentry.api.authentication import DSNAuthentication from sentry.api.base import EnvironmentMixin @@ -19,7 +21,7 @@ class Meta: class ProjectUserReportsEndpoint(ProjectEndpoint, EnvironmentMixin): authentication_classes = ProjectEndpoint.authentication_classes + (DSNAuthentication,) - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's User Feedback `````````````````````````````` @@ -65,7 +67,7 @@ def get(self, request, project): **paginate_kwargs, ) - def post(self, request, project): + def post(self, request: Request, project) -> Response: """ Submit User Feedback ```````````````````` diff --git a/src/sentry/api/endpoints/project_user_stats.py b/src/sentry/api/endpoints/project_user_stats.py index 1507651d37e32b..7a42fd599a1ddc 100644 --- a/src/sentry/api/endpoints/project_user_stats.py +++ b/src/sentry/api/endpoints/project_user_stats.py @@ -1,6 +1,7 @@ from datetime import timedelta from django.utils import timezone +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.base import EnvironmentMixin @@ -11,7 +12,7 @@ class ProjectUserStatsEndpoint(EnvironmentMixin, ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: try: environment_id = self._get_environment_id_from_request(request, project.organization_id) except Environment.DoesNotExist: diff --git a/src/sentry/api/endpoints/project_users.py b/src/sentry/api/endpoints/project_users.py index 7c5cdfa795c33a..0102c2146c1e5d 100644 --- a/src/sentry/api/endpoints/project_users.py +++ b/src/sentry/api/endpoints/project_users.py @@ -1,4 +1,5 @@ from rest_framework import status +from rest_framework.request import Request from rest_framework.response import Response from sentry.api.bases.project import ProjectEndpoint @@ -8,7 +9,7 @@ class ProjectUsersEndpoint(ProjectEndpoint): - def get(self, request, project): + def get(self, request: Request, project) -> Response: """ List a Project's Users ``````````````````````