mirror of
https://gitlab.com/gitlab-org/gitlab-foss.git
synced 2025-07-21 23:43:41 +00:00
1362 lines
42 KiB
Ruby
1362 lines
42 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
module Ci
|
|
class Build < Ci::Processable
|
|
prepend Ci::BulkInsertableTags
|
|
include Ci::Contextable
|
|
include Ci::Deployable
|
|
include TokenAuthenticatable
|
|
include AfterCommitQueue
|
|
include Presentable
|
|
include Importable
|
|
include Ci::HasRef
|
|
include Ci::TrackEnvironmentUsage
|
|
include EachBatch
|
|
include Ci::Taggable
|
|
|
|
extend ::Gitlab::Utils::Override
|
|
|
|
self.allow_legacy_sti_class = true
|
|
|
|
belongs_to :project, inverse_of: :builds
|
|
belongs_to :runner
|
|
belongs_to :erased_by, class_name: 'User'
|
|
belongs_to :pipeline,
|
|
->(build) { in_partition(build) },
|
|
class_name: 'Ci::Pipeline',
|
|
foreign_key: :commit_id,
|
|
partition_foreign_key: :partition_id,
|
|
inverse_of: :builds
|
|
belongs_to :project_mirror, primary_key: :project_id, foreign_key: :project_id, inverse_of: :builds
|
|
|
|
belongs_to :execution_config,
|
|
->(build) { in_partition(build) },
|
|
class_name: 'Ci::BuildExecutionConfig',
|
|
foreign_key: :execution_config_id,
|
|
partition_foreign_key: :partition_id,
|
|
inverse_of: :builds
|
|
|
|
RUNNER_FEATURES = {
|
|
upload_multiple_artifacts: ->(build) { build.publishes_artifacts_reports? },
|
|
refspecs: ->(build) { build.merge_request_ref? },
|
|
artifacts_exclude: ->(build) { build.supports_artifacts_exclude? },
|
|
multi_build_steps: ->(build) { build.multi_build_steps? },
|
|
return_exit_code: ->(build) { build.exit_codes_defined? },
|
|
fallback_cache_keys: ->(build) { build.fallback_cache_keys_defined? }
|
|
}.freeze
|
|
|
|
DEGRADATION_THRESHOLD_VARIABLE_NAME = 'DEGRADATION_THRESHOLD'
|
|
RUNNERS_STATUS_CACHE_EXPIRATION = 1.minute
|
|
DEPLOYMENT_NAMES = %w[deploy release rollout].freeze
|
|
|
|
TOKEN_PREFIX = 'glcbt-'
|
|
|
|
has_one :pending_state, class_name: 'Ci::BuildPendingState', foreign_key: :build_id, inverse_of: :build
|
|
has_one :queuing_entry, class_name: 'Ci::PendingBuild', foreign_key: :build_id, inverse_of: :build
|
|
has_one :runtime_metadata, class_name: 'Ci::RunningBuild', foreign_key: :build_id, inverse_of: :build
|
|
has_many :trace_chunks,
|
|
->(build) { in_partition(build) },
|
|
class_name: 'Ci::BuildTraceChunk',
|
|
foreign_key: :build_id,
|
|
inverse_of: :build,
|
|
partition_foreign_key: :partition_id
|
|
has_many :report_results, class_name: 'Ci::BuildReportResult', foreign_key: :build_id, inverse_of: :build
|
|
has_one :namespace, through: :project
|
|
|
|
has_one :build_source,
|
|
->(build) { in_partition(build) },
|
|
class_name: 'Ci::BuildSource',
|
|
foreign_key: :build_id,
|
|
inverse_of: :build,
|
|
partition_foreign_key: :partition_id
|
|
|
|
# Projects::DestroyService destroys Ci::Pipelines, which use_fast_destroy on :job_artifacts
|
|
# before we delete builds. By doing this, the relation should be empty and not fire any
|
|
# DELETE queries when the Ci::Build is destroyed. The next step is to remove `dependent: :destroy`.
|
|
# Details: https://gitlab.com/gitlab-org/gitlab/-/issues/24644#note_689472685
|
|
# rubocop:disable Cop/ActiveRecordDependent -- See above
|
|
has_many :job_artifacts,
|
|
->(build) { in_partition(build) },
|
|
class_name: 'Ci::JobArtifact',
|
|
foreign_key: :job_id,
|
|
partition_foreign_key: :partition_id,
|
|
dependent: :destroy,
|
|
inverse_of: :job
|
|
# rubocop:enable Cop/ActiveRecordDependent
|
|
|
|
has_many :job_variables, class_name: 'Ci::JobVariable', foreign_key: :job_id, inverse_of: :job
|
|
has_many :job_annotations,
|
|
->(build) { in_partition(build) },
|
|
class_name: 'Ci::JobAnnotation',
|
|
foreign_key: :job_id,
|
|
partition_foreign_key: :partition_id,
|
|
inverse_of: :job
|
|
has_many :sourced_pipelines, class_name: 'Ci::Sources::Pipeline', foreign_key: :source_job_id, inverse_of: :build
|
|
|
|
has_many :pages_deployments, foreign_key: :ci_build_id, inverse_of: :ci_build
|
|
|
|
has_many :taggings, ->(build) { in_partition(build) },
|
|
class_name: 'Ci::BuildTag',
|
|
foreign_key: :build_id,
|
|
partition_foreign_key: :partition_id,
|
|
inverse_of: :build
|
|
|
|
has_many :tags,
|
|
class_name: 'Ci::Tag',
|
|
through: :taggings,
|
|
source: :tag
|
|
|
|
Ci::JobArtifact.file_types.each_key do |key|
|
|
has_one :"job_artifacts_#{key}", ->(build) { in_partition(build).with_file_types([key]) },
|
|
class_name: 'Ci::JobArtifact',
|
|
foreign_key: :job_id,
|
|
partition_foreign_key: :partition_id,
|
|
inverse_of: :job
|
|
end
|
|
|
|
has_one :runner_manager_build,
|
|
->(build) { in_partition(build) },
|
|
class_name: 'Ci::RunnerManagerBuild',
|
|
foreign_key: :build_id,
|
|
inverse_of: :build,
|
|
partition_foreign_key: :partition_id,
|
|
autosave: true
|
|
has_one :runner_manager, foreign_key: :runner_machine_id, through: :runner_manager_build, class_name: 'Ci::RunnerManager'
|
|
|
|
has_one :runner_session, class_name: 'Ci::BuildRunnerSession', validate: true, foreign_key: :build_id, inverse_of: :build
|
|
has_one :trace_metadata, class_name: 'Ci::BuildTraceMetadata', foreign_key: :build_id, inverse_of: :build
|
|
|
|
has_many :terraform_state_versions, class_name: 'Terraform::StateVersion', foreign_key: :ci_build_id, inverse_of: :build
|
|
|
|
accepts_nested_attributes_for :runner_session, update_only: true
|
|
accepts_nested_attributes_for :job_variables
|
|
|
|
delegate :url, to: :runner_session, prefix: true, allow_nil: true
|
|
delegate :terminal_specification, to: :runner_session, allow_nil: true
|
|
delegate :service_specification, to: :runner_session, allow_nil: true
|
|
delegate :gitlab_deploy_token, to: :project
|
|
delegate :harbor_integration, to: :project
|
|
delegate :apple_app_store_integration, to: :project
|
|
delegate :google_play_integration, to: :project
|
|
delegate :diffblue_cover_integration, to: :project
|
|
delegate :ensure_persistent_ref, to: :pipeline
|
|
delegate :enable_debug_trace!, to: :metadata
|
|
|
|
serialize :options # rubocop:disable Cop/ActiveRecordSerialize
|
|
serialize :yaml_variables, Gitlab::Serializer::Ci::Variables # rubocop:disable Cop/ActiveRecordSerialize
|
|
|
|
delegate :name, to: :project, prefix: true
|
|
|
|
validates :coverage, numericality: true, allow_blank: true
|
|
validates :ref, presence: true
|
|
|
|
scope :unstarted, -> { where(runner_id: nil) }
|
|
scope :with_any_artifacts, -> { where_exists(Ci::JobArtifact.scoped_build) }
|
|
scope :with_downloadable_artifacts, -> { where_exists(Ci::JobArtifact.scoped_build.downloadable) }
|
|
scope :with_erasable_artifacts, -> { where_exists(Ci::JobArtifact.scoped_build.erasable) }
|
|
scope :with_existing_job_artifacts, ->(query) { where_exists(Ci::JobArtifact.scoped_build.erasable.merge(query)) }
|
|
scope :without_archived_trace, -> { where_not_exists(Ci::JobArtifact.scoped_build.trace) }
|
|
scope :with_artifacts, ->(artifact_scope) { with_existing_job_artifacts(artifact_scope).eager_load_job_artifacts }
|
|
|
|
scope :eager_load_job_artifacts, -> { includes(:job_artifacts) }
|
|
scope :eager_load_tags, -> { includes(:tags) }
|
|
scope :eager_load_for_archiving_trace, -> { preload(:project, :pending_state) }
|
|
scope :eager_load_for_api, -> do
|
|
preload(
|
|
:job_artifacts_archive, :ci_stage, :job_artifacts, :runner, :tags, :runner_manager, :metadata,
|
|
pipeline: :project,
|
|
user: [:user_preference, :user_detail, :followees]
|
|
)
|
|
end
|
|
|
|
scope :with_exposed_artifacts, -> do
|
|
joins(:metadata).merge(Ci::BuildMetadata.with_exposed_artifacts)
|
|
.includes(:metadata, :job_artifacts_metadata)
|
|
end
|
|
|
|
scope :with_artifacts_not_expired, -> { with_downloadable_artifacts.where('artifacts_expire_at IS NULL OR artifacts_expire_at > ?', Time.current) }
|
|
scope :with_pipeline_locked_artifacts, -> { joins(:pipeline).where('pipeline.locked': Ci::Pipeline.lockeds[:artifacts_locked]) }
|
|
scope :last_month, -> { where('created_at > ?', Date.today - 1.month) }
|
|
scope :scheduled_actions, -> { where(when: :delayed, status: COMPLETED_STATUSES + %i[scheduled]) }
|
|
scope :with_live_trace, -> { where_exists(Ci::BuildTraceChunk.scoped_build) }
|
|
scope :with_stale_live_trace, -> { with_live_trace.finished_before(12.hours.ago) }
|
|
scope :finished_before, ->(date) { finished.where('finished_at < ?', date) }
|
|
# WARNING: This scope could lead to performance implications for large size of tables `ci_builds` and ci_runners`.
|
|
# See https://gitlab.com/gitlab-org/gitlab/-/merge_requests/123131
|
|
scope :with_runner_type, ->(runner_type) { joins(:runner).where(runner: { runner_type: runner_type }) }
|
|
|
|
scope :belonging_to_runner_manager, ->(runner_machine_id) do
|
|
joins(:runner_manager_build).where(p_ci_runner_machine_builds: { runner_machine_id: runner_machine_id })
|
|
end
|
|
|
|
scope :with_secure_reports_from_config_options, ->(job_types) do
|
|
joins(:metadata).where("#{Ci::BuildMetadata.quoted_table_name}.config_options -> 'artifacts' -> 'reports' ?| array[:job_types]", job_types: job_types)
|
|
end
|
|
|
|
scope :with_coverage, -> { where.not(coverage: nil) }
|
|
scope :without_coverage, -> { where(coverage: nil) }
|
|
scope :with_coverage_regex, -> { where.not(coverage_regex: nil) }
|
|
scope :id_before, ->(id) { where(arel_table[:id].lt(id)) }
|
|
scope :id_after, ->(id) { where(arel_table[:id].gt(id)) }
|
|
|
|
scope :in_merge_request, ->(merge_request_id) do
|
|
joins(:pipeline).where(Ci::Pipeline.arel_table[:merge_request_id].eq(merge_request_id))
|
|
end
|
|
|
|
scope :with_job_artifacts, -> { joins(:job_artifacts) }
|
|
# the queries in the scope below are for the following cases,
|
|
# 1. builds may not have artifacts, still a valid dependency
|
|
# 2. build's artifacts belong to the same project, a valid dependency
|
|
# 3. build's artifacts from other projects, a valid dependency only if the artifact's accessibility is public
|
|
scope :builds_with_accessible_artifacts, ->(project_id) do
|
|
with_job_artifacts.where(job_artifacts: { job_id: nil })
|
|
.or(with_job_artifacts.where(job_artifacts: { file_type: 'dotenv', accessibility: 'public' }))
|
|
.or(with_job_artifacts.where(project_id: project_id, job_artifacts: { file_type: 'dotenv' })).distinct
|
|
end
|
|
|
|
scope :with_pipeline_source_type, ->(pipeline_source_type) { joins(:pipeline).where(pipeline: { source: pipeline_source_type }) }
|
|
scope :created_after, ->(time) { where(arel_table[:created_at].gt(time)) }
|
|
scope :updated_after, ->(time) { where(arel_table[:updated_at].gt(time)) }
|
|
scope :for_project_ids, ->(project_ids) { where(project_id: project_ids) }
|
|
|
|
add_authentication_token_field :token,
|
|
encrypted: :required,
|
|
format_with_prefix: :prefix_and_partition_for_token
|
|
|
|
after_save :stick_build_if_status_changed
|
|
|
|
after_create unless: :importing? do |build|
|
|
run_after_commit { build.execute_hooks }
|
|
end
|
|
|
|
after_commit :track_ci_secrets_management_id_tokens_usage, on: :create, if: :id_tokens?
|
|
after_commit :track_ci_build_created_event, on: :create
|
|
|
|
class << self
|
|
# This is needed for url_for to work,
|
|
# as the controller is JobsController
|
|
def model_name
|
|
ActiveModel::Name.new(self, nil, 'job')
|
|
end
|
|
|
|
def with_preloads
|
|
preload(:job_artifacts_archive, :job_artifacts, :tags, project: [:namespace])
|
|
end
|
|
|
|
def clone_accessors
|
|
%i[pipeline project ref tag options name
|
|
allow_failure stage_idx trigger_request
|
|
yaml_variables when environment coverage_regex
|
|
description tag_list protected needs_attributes
|
|
job_variables_attributes resource_group scheduling_type
|
|
ci_stage partition_id id_tokens interruptible execution_config_id].freeze
|
|
end
|
|
|
|
def supported_keyset_orderings
|
|
{ id: [:desc] }
|
|
end
|
|
end
|
|
|
|
state_machine :status do
|
|
event :enqueue do
|
|
transition [:created, :skipped, :manual, :scheduled] => :preparing, if: :any_unmet_prerequisites?
|
|
end
|
|
|
|
event :enqueue_scheduled do
|
|
transition scheduled: :preparing, if: :any_unmet_prerequisites?
|
|
transition scheduled: :pending
|
|
end
|
|
|
|
event :enqueue_preparing do
|
|
transition preparing: :pending
|
|
end
|
|
|
|
event :actionize do
|
|
transition created: :manual
|
|
end
|
|
|
|
event :schedule do
|
|
transition created: :scheduled
|
|
end
|
|
|
|
event :unschedule do
|
|
transition scheduled: :manual
|
|
end
|
|
|
|
before_transition on: :enqueue_scheduled do |build|
|
|
build.scheduled_at.nil? || build.scheduled_at.past? # If false is returned, it stops the transition
|
|
end
|
|
|
|
before_transition scheduled: any do |build|
|
|
build.scheduled_at = nil
|
|
end
|
|
|
|
before_transition created: :scheduled do |build|
|
|
build.scheduled_at = build.options_scheduled_at
|
|
end
|
|
|
|
before_transition on: :enqueue_preparing do |build|
|
|
!build.any_unmet_prerequisites? # If false is returned, it stops the transition
|
|
end
|
|
|
|
before_transition any => [:pending] do |build|
|
|
build.ensure_token
|
|
true
|
|
end
|
|
|
|
after_transition created: :scheduled do |build|
|
|
build.run_after_commit do
|
|
Ci::BuildScheduleWorker.perform_at(build.scheduled_at, build.id)
|
|
end
|
|
end
|
|
|
|
after_transition any => [:preparing] do |build|
|
|
build.run_after_commit do
|
|
Ci::BuildPrepareWorker.perform_async(id)
|
|
end
|
|
end
|
|
|
|
# rubocop:disable CodeReuse/ServiceClass
|
|
after_transition any => [:pending] do |build, transition|
|
|
Ci::UpdateBuildQueueService.new.push(build, transition)
|
|
|
|
build.run_after_commit do
|
|
BuildQueueWorker.perform_async(id)
|
|
build.execute_hooks
|
|
end
|
|
end
|
|
|
|
after_transition pending: any do |build, transition|
|
|
Ci::UpdateBuildQueueService.new.pop(build, transition)
|
|
end
|
|
|
|
after_transition any => [:running] do |build, transition|
|
|
Ci::UpdateBuildQueueService.new.track(build, transition)
|
|
end
|
|
|
|
after_transition running: any do |build, transition|
|
|
Ci::UpdateBuildQueueService.new.untrack(build, transition)
|
|
|
|
Ci::BuildRunnerSession.where(build: build).delete_all
|
|
end
|
|
|
|
# rubocop:enable CodeReuse/ServiceClass
|
|
#
|
|
after_transition pending: :running do |build|
|
|
build.ensure_metadata.update_timeout_state
|
|
end
|
|
|
|
after_transition pending: :running do |build|
|
|
build.run_after_commit do
|
|
build.ensure_persistent_ref
|
|
|
|
build.execute_hooks
|
|
end
|
|
end
|
|
|
|
after_transition any => [:success, :failed, :canceled] do |build|
|
|
build.run_after_commit do
|
|
build.run_status_commit_hooks!
|
|
|
|
Ci::BuildFinishedWorker.perform_async(id)
|
|
|
|
observe_report_types
|
|
end
|
|
end
|
|
|
|
after_transition any => [:success] do |build|
|
|
build.run_after_commit do
|
|
PagesWorker.perform_async(:deploy, id) if build.pages_generator?
|
|
end
|
|
end
|
|
|
|
after_transition any => [:failed] do |build|
|
|
next unless build.project
|
|
|
|
build.run_after_commit do
|
|
if build.auto_retry_allowed?
|
|
begin
|
|
# rubocop: disable CodeReuse/ServiceClass -- https://gitlab.com/gitlab-org/gitlab/-/issues/494865
|
|
Ci::RetryJobService.new(build.project, build.user).execute(build)
|
|
# rubocop: enable CodeReuse/ServiceClass
|
|
rescue Gitlab::Access::AccessDeniedError => e
|
|
Gitlab::AppLogger.error "Unable to auto-retry job #{build.id}: #{e}"
|
|
end
|
|
end
|
|
end
|
|
end
|
|
end
|
|
|
|
def self.build_matchers(project)
|
|
unique_params = [
|
|
:protected,
|
|
Arel.sql("(#{arel_tag_names_array.to_sql})")
|
|
]
|
|
|
|
group(*unique_params).pluck('array_agg(id)', *unique_params).map do |values|
|
|
Gitlab::Ci::Matching::BuildMatcher.new({
|
|
build_ids: values[0],
|
|
protected: values[1],
|
|
tag_list: values[2],
|
|
project: project
|
|
})
|
|
end
|
|
end
|
|
|
|
def self.ids_in_merge_request(merge_request_id)
|
|
in_merge_request(merge_request_id).pluck(:id)
|
|
end
|
|
|
|
def self.taggings_join_model
|
|
::Ci::BuildTag
|
|
end
|
|
|
|
# A Ci::Bridge may transition to `canceling` as a result of strategy: :depend
|
|
# but only a Ci::Build will transition to `canceling`` via `.cancel`
|
|
def supports_canceling?
|
|
cancel_gracefully?
|
|
end
|
|
|
|
def cancel_gracefully?
|
|
!!runner_manager&.supports_after_script_on_cancel?
|
|
end
|
|
|
|
def supports_force_cancel?
|
|
true
|
|
end
|
|
|
|
def build_matcher
|
|
strong_memoize(:build_matcher) do
|
|
Gitlab::Ci::Matching::BuildMatcher.new({
|
|
protected: protected?,
|
|
tag_list: tag_list,
|
|
build_ids: [id],
|
|
project: project
|
|
})
|
|
end
|
|
end
|
|
|
|
def auto_retry_allowed?
|
|
auto_retry.allowed?
|
|
end
|
|
|
|
def exit_code=(value)
|
|
return unless value
|
|
|
|
ensure_metadata.exit_code = value.to_i.clamp(0, Gitlab::Database::MAX_SMALLINT_VALUE)
|
|
end
|
|
|
|
def auto_retry_expected?
|
|
failed? && auto_retry_allowed?
|
|
end
|
|
|
|
def detailed_status(current_user)
|
|
Gitlab::Ci::Status::Build::Factory
|
|
.new(present, current_user)
|
|
.fabricate!
|
|
end
|
|
|
|
def other_scheduled_actions
|
|
pipeline.scheduled_actions.reject { |action| action.name == name }
|
|
end
|
|
|
|
def pages_generator?
|
|
return false unless Gitlab.config.pages.enabled
|
|
return false unless options.present?
|
|
return true if options[:pages].is_a?(Hash) || options[:pages] == true
|
|
|
|
options[:pages] != false && name == 'pages' # Legacy behaviour
|
|
end
|
|
|
|
def pages
|
|
return {} unless pages_generator? && publish_path_available?
|
|
|
|
{ publish: expanded_publish_path }
|
|
end
|
|
strong_memoize_attr :pages
|
|
|
|
def runnable?
|
|
true
|
|
end
|
|
|
|
def degenerated?
|
|
super && execution_config_id.nil?
|
|
end
|
|
|
|
def degenerate!
|
|
super do
|
|
execution_config&.destroy
|
|
end
|
|
end
|
|
|
|
def archived?
|
|
degenerated? || super
|
|
end
|
|
|
|
def playable?
|
|
action? && !archived? && (manual? || scheduled? || retryable?)
|
|
end
|
|
|
|
def schedulable?
|
|
self.when == 'delayed' && options[:start_in].present?
|
|
end
|
|
|
|
def options_scheduled_at
|
|
ChronicDuration.parse(options[:start_in])&.seconds&.from_now
|
|
end
|
|
|
|
def action?
|
|
ACTIONABLE_WHEN.include?(self.when)
|
|
end
|
|
|
|
def can_auto_cancel_pipeline_on_job_failure?
|
|
# A job that doesn't need to be auto-retried can auto-cancel its own pipeline
|
|
!auto_retry_expected?
|
|
end
|
|
|
|
# rubocop: disable CodeReuse/ServiceClass
|
|
def play(current_user, job_variables_attributes = nil)
|
|
Ci::PlayBuildService.new(current_user: current_user, build: self, variables: job_variables_attributes).execute
|
|
end
|
|
# rubocop: enable CodeReuse/ServiceClass
|
|
|
|
def cancelable?
|
|
(active? || created?) && !canceling?
|
|
end
|
|
|
|
def force_cancelable?
|
|
canceling? && supports_force_cancel?
|
|
end
|
|
|
|
def retries_count
|
|
pipeline.builds.retried.where(name: name).count
|
|
end
|
|
|
|
override :all_met_to_become_pending?
|
|
def all_met_to_become_pending?
|
|
super && !any_unmet_prerequisites?
|
|
end
|
|
|
|
def any_unmet_prerequisites?
|
|
prerequisites.present?
|
|
end
|
|
|
|
def prerequisites
|
|
Gitlab::Ci::Build::Prerequisite::Factory.new(self).unmet
|
|
end
|
|
|
|
def triggered_by?(current_user)
|
|
user == current_user
|
|
end
|
|
|
|
##
|
|
# All variables, including persisted environment variables.
|
|
#
|
|
def variables
|
|
strong_memoize(:variables) do
|
|
Gitlab::Ci::Variables::Collection.new
|
|
.concat(base_variables)
|
|
.concat(pages_variables)
|
|
end
|
|
end
|
|
|
|
def persisted_variables
|
|
Gitlab::Ci::Variables::Collection.new.tap do |variables|
|
|
break variables unless persisted?
|
|
|
|
variables
|
|
.concat(pipeline.persisted_variables)
|
|
.append(key: 'CI_JOB_ID', value: id.to_s)
|
|
.append(key: 'CI_JOB_URL', value: Gitlab::Routing.url_helpers.project_job_url(project, self))
|
|
.append(key: 'CI_JOB_TOKEN', value: token.to_s, public: false, masked: true)
|
|
.append(key: 'CI_JOB_STARTED_AT', value: started_at&.iso8601)
|
|
.append(key: 'CI_REGISTRY_USER', value: ::Gitlab::Auth::CI_JOB_USER)
|
|
.append(key: 'CI_REGISTRY_PASSWORD', value: token.to_s, public: false, masked: true)
|
|
.append(key: 'CI_REPOSITORY_URL', value: repo_url.to_s, public: false)
|
|
.concat(deploy_token_variables)
|
|
.concat(harbor_variables)
|
|
.concat(apple_app_store_variables)
|
|
.concat(google_play_variables)
|
|
.concat(diffblue_cover_variables)
|
|
end
|
|
end
|
|
|
|
def persisted_environment_variables
|
|
Gitlab::Ci::Variables::Collection.new.tap do |variables|
|
|
break variables unless persisted? && persisted_environment.present?
|
|
|
|
variables.append(key: 'CI_ENVIRONMENT_SLUG', value: environment_slug)
|
|
|
|
# Here we're passing unexpanded environment_url for runner to expand,
|
|
# and we need to make sure that CI_ENVIRONMENT_NAME and
|
|
# CI_ENVIRONMENT_SLUG so on are available for the URL be expanded.
|
|
variables.append(key: 'CI_ENVIRONMENT_URL', value: environment_url) if environment_url
|
|
end
|
|
end
|
|
|
|
def deploy_token_variables
|
|
Gitlab::Ci::Variables::Collection.new.tap do |variables|
|
|
break variables unless gitlab_deploy_token
|
|
|
|
variables.append(key: 'CI_DEPLOY_USER', value: gitlab_deploy_token.username)
|
|
variables.append(key: 'CI_DEPLOY_PASSWORD', value: gitlab_deploy_token.token, public: false, masked: true)
|
|
end
|
|
end
|
|
|
|
def dependency_proxy_variables
|
|
Gitlab::Ci::Variables::Collection.new.tap do |variables|
|
|
break variables unless Gitlab.config.dependency_proxy.enabled
|
|
|
|
variables.append(key: 'CI_DEPENDENCY_PROXY_USER', value: ::Gitlab::Auth::CI_JOB_USER)
|
|
variables.append(key: 'CI_DEPENDENCY_PROXY_PASSWORD', value: token.to_s, public: false, masked: true)
|
|
end
|
|
end
|
|
|
|
def harbor_variables
|
|
return [] unless harbor_integration.try(:activated?)
|
|
|
|
Gitlab::Ci::Variables::Collection.new(harbor_integration.ci_variables)
|
|
end
|
|
|
|
def apple_app_store_variables
|
|
return [] unless apple_app_store_integration.try(:activated?)
|
|
|
|
Gitlab::Ci::Variables::Collection.new(apple_app_store_integration.ci_variables(protected_ref: pipeline.protected_ref?))
|
|
end
|
|
|
|
def google_play_variables
|
|
return [] unless google_play_integration.try(:activated?)
|
|
|
|
Gitlab::Ci::Variables::Collection.new(google_play_integration.ci_variables(protected_ref: pipeline.protected_ref?))
|
|
end
|
|
|
|
def diffblue_cover_variables
|
|
return [] unless diffblue_cover_integration.try(:activated?)
|
|
|
|
Gitlab::Ci::Variables::Collection.new(diffblue_cover_integration.ci_variables)
|
|
end
|
|
|
|
def pages_variables
|
|
::Gitlab::Ci::Variables::Collection.new.tap do |variables|
|
|
variables.append(key: 'CI_PAGES_URL', value: project.pages_url(pages))
|
|
end
|
|
end
|
|
|
|
def base_variables
|
|
::Gitlab::Ci::Variables::Collection.new
|
|
.concat(persisted_variables)
|
|
.concat(dependency_proxy_variables)
|
|
.concat(job_jwt_variables)
|
|
.concat(scoped_variables)
|
|
.concat(job_variables)
|
|
.concat(persisted_environment_variables)
|
|
end
|
|
strong_memoize_attr :base_variables
|
|
|
|
def features
|
|
{
|
|
trace_sections: true,
|
|
failure_reasons: self.class.failure_reasons.keys
|
|
}
|
|
end
|
|
|
|
def merge_request
|
|
strong_memoize(:merge_request) do
|
|
pipeline.all_merge_requests.order(iid: :asc).first
|
|
end
|
|
end
|
|
|
|
def repo_url
|
|
return unless token
|
|
|
|
auth = "#{::Gitlab::Auth::CI_JOB_USER}:#{token}@"
|
|
project.http_url_to_repo.sub(%r{^https?://}) do |prefix|
|
|
prefix + auth
|
|
end
|
|
end
|
|
|
|
def allow_git_fetch
|
|
project.build_allow_git_fetch
|
|
end
|
|
|
|
def update_coverage
|
|
coverage = trace.extract_coverage(coverage_regex)
|
|
update(coverage: coverage) if coverage.present?
|
|
end
|
|
|
|
def trace
|
|
Gitlab::Ci::Trace.new(self)
|
|
end
|
|
|
|
def has_trace?
|
|
trace.exist?
|
|
end
|
|
|
|
def has_live_trace?
|
|
trace.live?
|
|
end
|
|
|
|
def has_archived_trace?
|
|
trace.archived?
|
|
end
|
|
|
|
def artifacts_file
|
|
job_artifacts_archive&.file
|
|
end
|
|
|
|
def artifacts_size
|
|
job_artifacts_archive&.size
|
|
end
|
|
|
|
def artifacts_metadata
|
|
job_artifacts_metadata&.file
|
|
end
|
|
|
|
def artifacts?
|
|
!artifacts_expired? && artifacts_file&.exists?
|
|
end
|
|
|
|
def locked_artifacts?
|
|
pipeline.artifacts_locked? && artifacts_file&.exists?
|
|
end
|
|
|
|
# This method is similar to #artifacts? but it includes the artifacts
|
|
# locking mechanics. A new method was created to prevent breaking existing
|
|
# behavior and avoid introducing N+1s.
|
|
def available_artifacts?
|
|
(!artifacts_expired? || pipeline.artifacts_locked?) && job_artifacts_archive&.exists?
|
|
end
|
|
|
|
def artifacts_metadata?
|
|
artifacts? && artifacts_metadata&.exists?
|
|
end
|
|
|
|
def has_job_artifacts?
|
|
job_artifacts.any?
|
|
end
|
|
|
|
def has_test_reports?
|
|
job_artifacts.of_report_type(:test).exists?
|
|
end
|
|
|
|
def ensure_trace_metadata!
|
|
Ci::BuildTraceMetadata.find_or_upsert_for!(id, partition_id)
|
|
end
|
|
|
|
def artifacts_expose_as
|
|
options.dig(:artifacts, :expose_as)
|
|
end
|
|
|
|
def artifacts_paths
|
|
options.dig(:artifacts, :paths)
|
|
end
|
|
|
|
def needs_touch?
|
|
Time.current - updated_at > 15.minutes.to_i
|
|
end
|
|
|
|
def remove_token!
|
|
update!(token_encrypted: nil)
|
|
end
|
|
|
|
def has_tags?
|
|
tag_list.any?
|
|
end
|
|
|
|
def any_runners_online?
|
|
cache_for_online_runners do
|
|
project.any_online_runners? { |runner| runner.match_build_if_online?(self) }
|
|
end
|
|
end
|
|
|
|
def any_runners_available?
|
|
cache_for_available_runners do
|
|
project.active_runners.exists?
|
|
end
|
|
end
|
|
|
|
def stuck?
|
|
pending? && !any_runners_online?
|
|
end
|
|
|
|
def execute_hooks
|
|
return unless project
|
|
return if user&.blocked?
|
|
|
|
return unless project.has_active_hooks?(:job_hooks) || project.has_active_integrations?(:job_hooks)
|
|
|
|
Ci::ExecuteBuildHooksWorker.perform_async(project.id, build_data)
|
|
end
|
|
|
|
def browsable_artifacts?
|
|
artifacts_metadata?
|
|
end
|
|
|
|
def artifacts_public?
|
|
return true if job_artifacts_archive.nil? # To backward compatibility return true if no artifacts found
|
|
|
|
job_artifacts_archive.public_access?
|
|
end
|
|
|
|
def artifacts_no_access?
|
|
return false if job_artifacts_archive.nil? # To backward compatibility return false if no artifacts found
|
|
|
|
job_artifacts_archive.none_access?
|
|
end
|
|
|
|
def artifact_access_setting_in_config
|
|
artifacts_public = options.dig(:artifacts, :public)
|
|
artifacts_access = options.dig(:artifacts, :access)
|
|
|
|
if !artifacts_public.nil? && !artifacts_access.nil?
|
|
raise ArgumentError, 'artifacts:public and artifacts:access are mutually exclusive'
|
|
end
|
|
|
|
return :public if artifacts_public == true || artifacts_access == 'all'
|
|
return :private if artifacts_public == false || artifacts_access == 'developer'
|
|
return :none if artifacts_access == 'none'
|
|
|
|
# default behaviour
|
|
:public
|
|
end
|
|
|
|
def artifacts_metadata_entry(path, **options)
|
|
artifacts_metadata.open do |metadata_stream|
|
|
metadata = Gitlab::Ci::Build::Artifacts::Metadata.new(
|
|
metadata_stream,
|
|
path,
|
|
**options)
|
|
|
|
metadata.to_entry
|
|
end
|
|
end
|
|
|
|
def erasable?
|
|
complete? && (artifacts? || has_job_artifacts? || has_trace?)
|
|
end
|
|
|
|
def erased?
|
|
!erased_at.nil?
|
|
end
|
|
|
|
def artifacts_expired?
|
|
artifacts_expire_at&.past?
|
|
end
|
|
|
|
def artifacts_expire_in
|
|
artifacts_expire_at - Time.current if artifacts_expire_at
|
|
end
|
|
|
|
def artifacts_expire_in=(value)
|
|
self.artifacts_expire_at =
|
|
(ChronicDuration.parse(value)&.seconds&.from_now if value)
|
|
end
|
|
|
|
def has_expired_locked_archive_artifacts?
|
|
locked_artifacts? &&
|
|
artifacts_expire_at&.past?
|
|
end
|
|
|
|
def has_expiring_archive_artifacts?
|
|
has_expiring_artifacts? && job_artifacts_archive.present?
|
|
end
|
|
|
|
def self.keep_artifacts!
|
|
update_all(artifacts_expire_at: nil)
|
|
Ci::JobArtifact.where(job: self.select(:id)).update_all(expire_at: nil)
|
|
end
|
|
|
|
def keep_artifacts!
|
|
update(artifacts_expire_at: nil)
|
|
job_artifacts.update_all(expire_at: nil)
|
|
end
|
|
|
|
def artifact_for_type(type)
|
|
file_types = Ci::JobArtifact.associated_file_types_for(type)
|
|
file_types_ids = file_types&.map { |file_type| Ci::JobArtifact.file_types[file_type] }
|
|
job_artifacts.find_by(file_type: file_types_ids)
|
|
end
|
|
|
|
def steps
|
|
[Gitlab::Ci::Build::Step.from_commands(self),
|
|
Gitlab::Ci::Build::Step.from_release(self),
|
|
Gitlab::Ci::Build::Step.from_after_script(self)].compact
|
|
end
|
|
|
|
def runtime_hooks
|
|
Gitlab::Ci::Build::Hook.from_hooks(self)
|
|
end
|
|
|
|
def image
|
|
Gitlab::Ci::Build::Image.from_image(self)
|
|
end
|
|
|
|
def services
|
|
Gitlab::Ci::Build::Image.from_services(self)
|
|
end
|
|
|
|
def cache
|
|
cache = Array.wrap(options[:cache])
|
|
|
|
cache.each do |single_cache|
|
|
single_cache[:fallback_keys] = [] unless single_cache.key?(:fallback_keys)
|
|
end
|
|
|
|
if project.jobs_cache_index
|
|
cache = cache.map do |single_cache|
|
|
cache = single_cache.merge(key: "#{single_cache[:key]}-#{project.jobs_cache_index}")
|
|
fallback = cache.slice(:fallback_keys).transform_values { |keys| keys.map { |key| "#{key}-#{project.jobs_cache_index}" } }
|
|
cache.merge(fallback.compact)
|
|
end
|
|
end
|
|
|
|
return cache unless project.ci_separated_caches
|
|
|
|
cache.map do |entry|
|
|
type_suffix = !entry[:unprotect] && pipeline.protected_ref? ? 'protected' : 'non_protected'
|
|
|
|
cache = entry.merge(key: "#{entry[:key]}-#{type_suffix}")
|
|
fallback = cache.slice(:fallback_keys).transform_values { |keys| keys.map { |key| "#{key}-#{type_suffix}" } }
|
|
cache.merge(fallback.compact)
|
|
end
|
|
end
|
|
|
|
def fallback_cache_keys_defined?
|
|
Array.wrap(options[:cache]).any? { |cache| cache[:fallback_keys].present? }
|
|
end
|
|
|
|
def credentials
|
|
Gitlab::Ci::Build::Credentials::Factory.new(self).create!
|
|
end
|
|
|
|
def has_valid_build_dependencies?
|
|
dependencies.valid?
|
|
end
|
|
|
|
def invalid_dependencies
|
|
dependencies.invalid_local
|
|
end
|
|
|
|
def valid_dependency?
|
|
return false if artifacts_expired? && !pipeline.artifacts_locked?
|
|
return false if erased?
|
|
|
|
true
|
|
end
|
|
|
|
def runner_required_feature_names
|
|
strong_memoize(:runner_required_feature_names) do
|
|
RUNNER_FEATURES.select do |feature, method|
|
|
method.call(self)
|
|
end.keys
|
|
end
|
|
end
|
|
|
|
def supported_runner?(features)
|
|
runner_required_feature_names.all? do |feature_name|
|
|
features&.dig(feature_name)
|
|
end
|
|
end
|
|
|
|
def publishes_artifacts_reports?
|
|
options&.dig(:artifacts, :reports)&.any?
|
|
end
|
|
|
|
def supports_artifacts_exclude?
|
|
options&.dig(:artifacts, :exclude)&.any?
|
|
end
|
|
|
|
def publish_path
|
|
return unless options.present?
|
|
return options[:publish] unless options[:pages].is_a?(Hash)
|
|
|
|
options.dig(:pages, :publish) || options[:publish]
|
|
end
|
|
|
|
def publish_path_available?
|
|
publish_path.present?
|
|
end
|
|
|
|
def expanded_publish_path
|
|
ExpandVariables.expand(publish_path.to_s, -> { base_variables.sort_and_expand_all })
|
|
end
|
|
|
|
def multi_build_steps?
|
|
options[:release]&.any?
|
|
end
|
|
|
|
def hide_secrets(data, metrics = ::Gitlab::Ci::Trace::Metrics.new)
|
|
return unless trace
|
|
|
|
data.dup.tap do |trace|
|
|
Gitlab::Ci::MaskSecret.mask!(trace, project.runners_token) if project
|
|
Gitlab::Ci::MaskSecret.mask!(trace, token) if token
|
|
|
|
metrics.increment_trace_operation(operation: :mutated) if trace != data
|
|
end
|
|
end
|
|
|
|
def serializable_hash(options = {})
|
|
super(options).merge(when: read_attribute(:when))
|
|
end
|
|
|
|
def has_terminal?
|
|
running? && runner_session_url.present?
|
|
end
|
|
|
|
def collect_test_reports!(test_reports)
|
|
each_report(Ci::JobArtifact.file_types_for_report(:test)) do |file_type, blob|
|
|
Gitlab::Ci::Parsers.fabricate!(file_type).parse!(blob, test_reports, job: self)
|
|
end
|
|
|
|
test_reports
|
|
end
|
|
|
|
def collect_accessibility_reports!(accessibility_report)
|
|
each_report(Ci::JobArtifact.file_types_for_report(:accessibility)) do |file_type, blob|
|
|
Gitlab::Ci::Parsers.fabricate!(file_type).parse!(blob, accessibility_report)
|
|
end
|
|
|
|
accessibility_report
|
|
end
|
|
|
|
def collect_codequality_reports!(codequality_report)
|
|
each_report(Ci::JobArtifact.file_types_for_report(:codequality)) do |file_type, blob|
|
|
Gitlab::Ci::Parsers.fabricate!(file_type).parse!(blob, codequality_report, { project: project, commit_sha: pipeline.sha })
|
|
end
|
|
|
|
codequality_report
|
|
end
|
|
|
|
def collect_terraform_reports!(terraform_reports)
|
|
each_report(::Ci::JobArtifact.file_types_for_report(:terraform)) do |file_type, blob, report_artifact|
|
|
::Gitlab::Ci::Parsers.fabricate!(file_type).parse!(blob, terraform_reports, artifact: report_artifact)
|
|
end
|
|
|
|
terraform_reports
|
|
end
|
|
|
|
def report_artifacts
|
|
job_artifacts.all_reports
|
|
end
|
|
|
|
# Consider this object to have an unknown job problem
|
|
def doom!
|
|
transaction do
|
|
now = Time.current
|
|
attributes = {
|
|
status: :failed,
|
|
failure_reason: :data_integrity_failure,
|
|
updated_at: now
|
|
}
|
|
attributes[:finished_at] = now unless finished_at.present?
|
|
|
|
update_columns(attributes)
|
|
all_queuing_entries.delete_all
|
|
all_runtime_metadata.delete_all
|
|
end
|
|
|
|
deployment&.sync_status_with(self)
|
|
|
|
::Gitlab::Ci::Pipeline::Metrics
|
|
.job_failure_reason_counter
|
|
.increment(reason: :data_integrity_failure)
|
|
|
|
Gitlab::AppLogger.info(
|
|
message: 'Build doomed',
|
|
class: self.class.name,
|
|
build_id: id,
|
|
pipeline_id: pipeline_id,
|
|
project_id: project_id)
|
|
end
|
|
|
|
def degradation_threshold
|
|
var = yaml_variables.find { |v| v[:key] == DEGRADATION_THRESHOLD_VARIABLE_NAME }
|
|
var[:value]&.to_i if var
|
|
end
|
|
|
|
def remove_pending_state!
|
|
pending_state.try(:delete)
|
|
end
|
|
|
|
def run_on_status_commit(&block)
|
|
status_commit_hooks.push(block)
|
|
end
|
|
|
|
def max_test_cases_per_report
|
|
# NOTE: This is temporary and will be replaced later by a value
|
|
# that would come from an actual application limit.
|
|
::Gitlab.com? ? 500_000 : 0
|
|
end
|
|
|
|
def debug_mode?
|
|
# perform the check on both sides in case the runner version is old
|
|
metadata&.debug_trace_enabled? ||
|
|
Gitlab::Utils.to_boolean(variables['CI_DEBUG_SERVICES']&.value, default: false) ||
|
|
Gitlab::Utils.to_boolean(variables['CI_DEBUG_TRACE']&.value, default: false)
|
|
end
|
|
|
|
def drop_with_exit_code!(failure_reason, exit_code)
|
|
failure_reason ||= :unknown_failure
|
|
result = drop!(::Gitlab::Ci::Build::Status::Reason.new(self, failure_reason, exit_code))
|
|
::Ci::TrackFailedBuildWorker.perform_async(id, exit_code, failure_reason)
|
|
result
|
|
end
|
|
|
|
def exit_codes_defined?
|
|
options.dig(:allow_failure_criteria, :exit_codes).present? || options.dig(:retry, :exit_codes).present?
|
|
end
|
|
|
|
def create_queuing_entry!
|
|
::Ci::PendingBuild.upsert_from_build!(self)
|
|
end
|
|
|
|
##
|
|
# We can have only one queuing entry or running build tracking entry,
|
|
# because there is a unique index on `build_id` in each table, but we need
|
|
# a relation to remove these entries more efficiently in a single statement
|
|
# without actually loading data.
|
|
#
|
|
def all_queuing_entries
|
|
::Ci::PendingBuild.where(build_id: id)
|
|
end
|
|
|
|
def all_runtime_metadata
|
|
::Ci::RunningBuild.where(build_id: id)
|
|
end
|
|
|
|
def shared_runner_build?
|
|
runner&.instance_type?
|
|
end
|
|
|
|
def job_variables_attributes
|
|
strong_memoize(:job_variables_attributes) do
|
|
job_variables.internal_source.map do |variable|
|
|
variable.attributes.except('id', 'job_id', 'encrypted_value', 'encrypted_value_iv').tap do |attrs|
|
|
attrs[:value] = variable.value
|
|
end
|
|
end
|
|
end
|
|
end
|
|
|
|
def allowed_to_fail_with_code?(exit_code)
|
|
options
|
|
.dig(:allow_failure_criteria, :exit_codes)
|
|
.to_a
|
|
.include?(exit_code)
|
|
end
|
|
|
|
def each_report(report_types)
|
|
job_artifacts_for_types(report_types).each do |report_artifact|
|
|
next if report_artifact&.artifact_report&.faulty?
|
|
|
|
report_artifact.each_blob do |blob|
|
|
yield report_artifact.file_type, blob, report_artifact
|
|
end
|
|
end
|
|
end
|
|
|
|
def clone(current_user:, new_job_variables_attributes: [])
|
|
new_build = super
|
|
|
|
if action? && new_job_variables_attributes.any?
|
|
new_build.job_variables = []
|
|
new_build.job_variables_attributes = new_job_variables_attributes
|
|
end
|
|
|
|
new_build
|
|
end
|
|
|
|
def job_artifact_types
|
|
job_artifacts.map(&:file_type)
|
|
end
|
|
|
|
def test_suite_name
|
|
if matrix_build?
|
|
name
|
|
else
|
|
group_name
|
|
end
|
|
end
|
|
|
|
def time_in_queue_seconds
|
|
return if queued_at.nil?
|
|
|
|
(::Time.current - queued_at).seconds.to_i
|
|
end
|
|
strong_memoize_attr :time_in_queue_seconds
|
|
|
|
def source
|
|
build_source&.source || pipeline.source
|
|
end
|
|
strong_memoize_attr :source
|
|
|
|
# Can be removed in Rails 7.1. Related to: Gitlab.next_rails?
|
|
def to_partial_path
|
|
'jobs/job'
|
|
end
|
|
|
|
def token
|
|
return encoded_jwt if user&.has_composite_identity? || use_jwt_for_ci_cd_job_token?
|
|
|
|
super
|
|
end
|
|
|
|
protected
|
|
|
|
def run_status_commit_hooks!
|
|
status_commit_hooks.reverse_each do |hook|
|
|
instance_eval(&hook)
|
|
end
|
|
end
|
|
|
|
private
|
|
|
|
def use_jwt_for_ci_cd_job_token?
|
|
namespace&.root_ancestor&.namespace_settings&.jwt_ci_cd_job_token_enabled?
|
|
end
|
|
|
|
def encoded_jwt
|
|
::Ci::JobToken::Jwt.encode(self)
|
|
end
|
|
strong_memoize_attr :encoded_jwt
|
|
|
|
def matrix_build?
|
|
options.dig(:parallel, :matrix).present?
|
|
end
|
|
|
|
def stick_build_if_status_changed
|
|
return unless saved_change_to_status?
|
|
return unless running?
|
|
|
|
self.class.sticking.stick(:build, id)
|
|
end
|
|
|
|
def status_commit_hooks
|
|
@status_commit_hooks ||= []
|
|
end
|
|
|
|
def auto_retry
|
|
strong_memoize(:auto_retry) do
|
|
Gitlab::Ci::Build::AutoRetry.new(self)
|
|
end
|
|
end
|
|
|
|
def build_data
|
|
strong_memoize(:build_data) do
|
|
ActiveRecord::Associations::Preloader.new(records: [self], associations: { runner: :tags }).call
|
|
Gitlab::DataBuilder::Build.build(self)
|
|
end
|
|
end
|
|
|
|
def job_artifacts_for_types(report_types)
|
|
# Use select to leverage cached associations and avoid N+1 queries
|
|
job_artifacts.select { |artifact| artifact.file_type.in?(report_types) }
|
|
end
|
|
|
|
def has_expiring_artifacts?
|
|
artifacts_expire_at.present? && artifacts_expire_at.future?
|
|
end
|
|
|
|
def job_jwt_variables
|
|
id_tokens_variables
|
|
end
|
|
|
|
def id_tokens_variables
|
|
Gitlab::Ci::Variables::Collection.new.tap do |variables|
|
|
break variables unless id_tokens?
|
|
|
|
sub_components = project.ci_id_token_sub_claim_components.map(&:to_sym)
|
|
|
|
id_tokens.each do |var_name, token_data|
|
|
token = Gitlab::Ci::JwtV2.for_build(self, aud: expanded_id_token_aud(token_data['aud']),
|
|
sub_components: sub_components)
|
|
|
|
variables.append(key: var_name, value: token, public: false, masked: true)
|
|
end
|
|
rescue OpenSSL::PKey::RSAError, Gitlab::Ci::Jwt::NoSigningKeyError => e
|
|
Gitlab::ErrorTracking.track_exception(e)
|
|
end
|
|
end
|
|
|
|
def expanded_id_token_aud(aud)
|
|
return unless aud
|
|
|
|
strong_memoize_with(:expanded_id_token_aud, aud) do
|
|
# `aud` can be a string or an array of strings.
|
|
if aud.is_a?(Array)
|
|
aud.map { |x| ExpandVariables.expand(x, -> { scoped_variables.sort_and_expand_all }) }
|
|
else
|
|
ExpandVariables.expand(aud, -> { scoped_variables.sort_and_expand_all })
|
|
end
|
|
end
|
|
end
|
|
|
|
def cache_for_online_runners(&block)
|
|
Rails.cache.fetch(
|
|
['has-online-runners', id],
|
|
expires_in: RUNNERS_STATUS_CACHE_EXPIRATION
|
|
) { yield }
|
|
end
|
|
|
|
def cache_for_available_runners(&block)
|
|
Rails.cache.fetch(
|
|
['has-available-runners', project.id],
|
|
expires_in: RUNNERS_STATUS_CACHE_EXPIRATION
|
|
) { yield }
|
|
end
|
|
|
|
def observe_report_types
|
|
return unless ::Gitlab.com?
|
|
|
|
report_types = options&.dig(:artifacts, :reports)&.keys || []
|
|
|
|
report_types.each do |report_type|
|
|
next unless Enums::Ci::JobArtifact.report_types.include?(report_type)
|
|
|
|
::Gitlab::Ci::Artifacts::Metrics
|
|
.build_completed_report_type_counter(report_type)
|
|
.increment(status: status)
|
|
end
|
|
end
|
|
|
|
def track_ci_secrets_management_id_tokens_usage
|
|
::Gitlab::UsageDataCounters::HLLRedisCounter.track_event('i_ci_secrets_management_id_tokens_build_created', values: user_id)
|
|
|
|
Gitlab::Tracking.event(
|
|
self.class.to_s,
|
|
'create_id_tokens',
|
|
namespace: namespace,
|
|
user: user,
|
|
label: 'redis_hll_counters.ci_secrets_management.i_ci_secrets_management_id_tokens_build_created_monthly',
|
|
ultimate_namespace_id: namespace.root_ancestor.id,
|
|
context: [Gitlab::Tracking::ServicePingContext.new(
|
|
data_source: :redis_hll,
|
|
event: 'i_ci_secrets_management_id_tokens_build_created'
|
|
).to_context]
|
|
)
|
|
end
|
|
|
|
def track_ci_build_created_event
|
|
Gitlab::InternalEvents.track_event(
|
|
'create_ci_build',
|
|
project: project,
|
|
user: user,
|
|
additional_properties: {
|
|
property: name
|
|
}
|
|
)
|
|
end
|
|
|
|
def partition_id_prefix_in_16_bit_encode
|
|
"#{partition_id.to_s(16)}_"
|
|
end
|
|
|
|
def prefix_and_partition_for_token
|
|
TOKEN_PREFIX + partition_id_prefix_in_16_bit_encode
|
|
end
|
|
end
|
|
end
|
|
|
|
Ci::Build.prepend_mod_with('Ci::Build')
|