Files
gitlab-foss/app/models/ci/build.rb
2025-04-19 00:08:58 +00:00

1362 lines
42 KiB
Ruby

# frozen_string_literal: true
module Ci
class Build < Ci::Processable
prepend Ci::BulkInsertableTags
include Ci::Contextable
include Ci::Deployable
include TokenAuthenticatable
include AfterCommitQueue
include Presentable
include Importable
include Ci::HasRef
include Ci::TrackEnvironmentUsage
include EachBatch
include Ci::Taggable
extend ::Gitlab::Utils::Override
self.allow_legacy_sti_class = true
belongs_to :project, inverse_of: :builds
belongs_to :runner
belongs_to :erased_by, class_name: 'User'
belongs_to :pipeline,
->(build) { in_partition(build) },
class_name: 'Ci::Pipeline',
foreign_key: :commit_id,
partition_foreign_key: :partition_id,
inverse_of: :builds
belongs_to :project_mirror, primary_key: :project_id, foreign_key: :project_id, inverse_of: :builds
belongs_to :execution_config,
->(build) { in_partition(build) },
class_name: 'Ci::BuildExecutionConfig',
foreign_key: :execution_config_id,
partition_foreign_key: :partition_id,
inverse_of: :builds
RUNNER_FEATURES = {
upload_multiple_artifacts: ->(build) { build.publishes_artifacts_reports? },
refspecs: ->(build) { build.merge_request_ref? },
artifacts_exclude: ->(build) { build.supports_artifacts_exclude? },
multi_build_steps: ->(build) { build.multi_build_steps? },
return_exit_code: ->(build) { build.exit_codes_defined? },
fallback_cache_keys: ->(build) { build.fallback_cache_keys_defined? }
}.freeze
DEGRADATION_THRESHOLD_VARIABLE_NAME = 'DEGRADATION_THRESHOLD'
RUNNERS_STATUS_CACHE_EXPIRATION = 1.minute
DEPLOYMENT_NAMES = %w[deploy release rollout].freeze
TOKEN_PREFIX = 'glcbt-'
has_one :pending_state, class_name: 'Ci::BuildPendingState', foreign_key: :build_id, inverse_of: :build
has_one :queuing_entry, class_name: 'Ci::PendingBuild', foreign_key: :build_id, inverse_of: :build
has_one :runtime_metadata, class_name: 'Ci::RunningBuild', foreign_key: :build_id, inverse_of: :build
has_many :trace_chunks,
->(build) { in_partition(build) },
class_name: 'Ci::BuildTraceChunk',
foreign_key: :build_id,
inverse_of: :build,
partition_foreign_key: :partition_id
has_many :report_results, class_name: 'Ci::BuildReportResult', foreign_key: :build_id, inverse_of: :build
has_one :namespace, through: :project
has_one :build_source,
->(build) { in_partition(build) },
class_name: 'Ci::BuildSource',
foreign_key: :build_id,
inverse_of: :build,
partition_foreign_key: :partition_id
# Projects::DestroyService destroys Ci::Pipelines, which use_fast_destroy on :job_artifacts
# before we delete builds. By doing this, the relation should be empty and not fire any
# DELETE queries when the Ci::Build is destroyed. The next step is to remove `dependent: :destroy`.
# Details: https://gitlab.com/gitlab-org/gitlab/-/issues/24644#note_689472685
# rubocop:disable Cop/ActiveRecordDependent -- See above
has_many :job_artifacts,
->(build) { in_partition(build) },
class_name: 'Ci::JobArtifact',
foreign_key: :job_id,
partition_foreign_key: :partition_id,
dependent: :destroy,
inverse_of: :job
# rubocop:enable Cop/ActiveRecordDependent
has_many :job_variables, class_name: 'Ci::JobVariable', foreign_key: :job_id, inverse_of: :job
has_many :job_annotations,
->(build) { in_partition(build) },
class_name: 'Ci::JobAnnotation',
foreign_key: :job_id,
partition_foreign_key: :partition_id,
inverse_of: :job
has_many :sourced_pipelines, class_name: 'Ci::Sources::Pipeline', foreign_key: :source_job_id, inverse_of: :build
has_many :pages_deployments, foreign_key: :ci_build_id, inverse_of: :ci_build
has_many :taggings, ->(build) { in_partition(build) },
class_name: 'Ci::BuildTag',
foreign_key: :build_id,
partition_foreign_key: :partition_id,
inverse_of: :build
has_many :tags,
class_name: 'Ci::Tag',
through: :taggings,
source: :tag
Ci::JobArtifact.file_types.each_key do |key|
has_one :"job_artifacts_#{key}", ->(build) { in_partition(build).with_file_types([key]) },
class_name: 'Ci::JobArtifact',
foreign_key: :job_id,
partition_foreign_key: :partition_id,
inverse_of: :job
end
has_one :runner_manager_build,
->(build) { in_partition(build) },
class_name: 'Ci::RunnerManagerBuild',
foreign_key: :build_id,
inverse_of: :build,
partition_foreign_key: :partition_id,
autosave: true
has_one :runner_manager, foreign_key: :runner_machine_id, through: :runner_manager_build, class_name: 'Ci::RunnerManager'
has_one :runner_session, class_name: 'Ci::BuildRunnerSession', validate: true, foreign_key: :build_id, inverse_of: :build
has_one :trace_metadata, class_name: 'Ci::BuildTraceMetadata', foreign_key: :build_id, inverse_of: :build
has_many :terraform_state_versions, class_name: 'Terraform::StateVersion', foreign_key: :ci_build_id, inverse_of: :build
accepts_nested_attributes_for :runner_session, update_only: true
accepts_nested_attributes_for :job_variables
delegate :url, to: :runner_session, prefix: true, allow_nil: true
delegate :terminal_specification, to: :runner_session, allow_nil: true
delegate :service_specification, to: :runner_session, allow_nil: true
delegate :gitlab_deploy_token, to: :project
delegate :harbor_integration, to: :project
delegate :apple_app_store_integration, to: :project
delegate :google_play_integration, to: :project
delegate :diffblue_cover_integration, to: :project
delegate :ensure_persistent_ref, to: :pipeline
delegate :enable_debug_trace!, to: :metadata
serialize :options # rubocop:disable Cop/ActiveRecordSerialize
serialize :yaml_variables, Gitlab::Serializer::Ci::Variables # rubocop:disable Cop/ActiveRecordSerialize
delegate :name, to: :project, prefix: true
validates :coverage, numericality: true, allow_blank: true
validates :ref, presence: true
scope :unstarted, -> { where(runner_id: nil) }
scope :with_any_artifacts, -> { where_exists(Ci::JobArtifact.scoped_build) }
scope :with_downloadable_artifacts, -> { where_exists(Ci::JobArtifact.scoped_build.downloadable) }
scope :with_erasable_artifacts, -> { where_exists(Ci::JobArtifact.scoped_build.erasable) }
scope :with_existing_job_artifacts, ->(query) { where_exists(Ci::JobArtifact.scoped_build.erasable.merge(query)) }
scope :without_archived_trace, -> { where_not_exists(Ci::JobArtifact.scoped_build.trace) }
scope :with_artifacts, ->(artifact_scope) { with_existing_job_artifacts(artifact_scope).eager_load_job_artifacts }
scope :eager_load_job_artifacts, -> { includes(:job_artifacts) }
scope :eager_load_tags, -> { includes(:tags) }
scope :eager_load_for_archiving_trace, -> { preload(:project, :pending_state) }
scope :eager_load_for_api, -> do
preload(
:job_artifacts_archive, :ci_stage, :job_artifacts, :runner, :tags, :runner_manager, :metadata,
pipeline: :project,
user: [:user_preference, :user_detail, :followees]
)
end
scope :with_exposed_artifacts, -> do
joins(:metadata).merge(Ci::BuildMetadata.with_exposed_artifacts)
.includes(:metadata, :job_artifacts_metadata)
end
scope :with_artifacts_not_expired, -> { with_downloadable_artifacts.where('artifacts_expire_at IS NULL OR artifacts_expire_at > ?', Time.current) }
scope :with_pipeline_locked_artifacts, -> { joins(:pipeline).where('pipeline.locked': Ci::Pipeline.lockeds[:artifacts_locked]) }
scope :last_month, -> { where('created_at > ?', Date.today - 1.month) }
scope :scheduled_actions, -> { where(when: :delayed, status: COMPLETED_STATUSES + %i[scheduled]) }
scope :with_live_trace, -> { where_exists(Ci::BuildTraceChunk.scoped_build) }
scope :with_stale_live_trace, -> { with_live_trace.finished_before(12.hours.ago) }
scope :finished_before, ->(date) { finished.where('finished_at < ?', date) }
# WARNING: This scope could lead to performance implications for large size of tables `ci_builds` and ci_runners`.
# See https://gitlab.com/gitlab-org/gitlab/-/merge_requests/123131
scope :with_runner_type, ->(runner_type) { joins(:runner).where(runner: { runner_type: runner_type }) }
scope :belonging_to_runner_manager, ->(runner_machine_id) do
joins(:runner_manager_build).where(p_ci_runner_machine_builds: { runner_machine_id: runner_machine_id })
end
scope :with_secure_reports_from_config_options, ->(job_types) do
joins(:metadata).where("#{Ci::BuildMetadata.quoted_table_name}.config_options -> 'artifacts' -> 'reports' ?| array[:job_types]", job_types: job_types)
end
scope :with_coverage, -> { where.not(coverage: nil) }
scope :without_coverage, -> { where(coverage: nil) }
scope :with_coverage_regex, -> { where.not(coverage_regex: nil) }
scope :id_before, ->(id) { where(arel_table[:id].lt(id)) }
scope :id_after, ->(id) { where(arel_table[:id].gt(id)) }
scope :in_merge_request, ->(merge_request_id) do
joins(:pipeline).where(Ci::Pipeline.arel_table[:merge_request_id].eq(merge_request_id))
end
scope :with_job_artifacts, -> { joins(:job_artifacts) }
# the queries in the scope below are for the following cases,
# 1. builds may not have artifacts, still a valid dependency
# 2. build's artifacts belong to the same project, a valid dependency
# 3. build's artifacts from other projects, a valid dependency only if the artifact's accessibility is public
scope :builds_with_accessible_artifacts, ->(project_id) do
with_job_artifacts.where(job_artifacts: { job_id: nil })
.or(with_job_artifacts.where(job_artifacts: { file_type: 'dotenv', accessibility: 'public' }))
.or(with_job_artifacts.where(project_id: project_id, job_artifacts: { file_type: 'dotenv' })).distinct
end
scope :with_pipeline_source_type, ->(pipeline_source_type) { joins(:pipeline).where(pipeline: { source: pipeline_source_type }) }
scope :created_after, ->(time) { where(arel_table[:created_at].gt(time)) }
scope :updated_after, ->(time) { where(arel_table[:updated_at].gt(time)) }
scope :for_project_ids, ->(project_ids) { where(project_id: project_ids) }
add_authentication_token_field :token,
encrypted: :required,
format_with_prefix: :prefix_and_partition_for_token
after_save :stick_build_if_status_changed
after_create unless: :importing? do |build|
run_after_commit { build.execute_hooks }
end
after_commit :track_ci_secrets_management_id_tokens_usage, on: :create, if: :id_tokens?
after_commit :track_ci_build_created_event, on: :create
class << self
# This is needed for url_for to work,
# as the controller is JobsController
def model_name
ActiveModel::Name.new(self, nil, 'job')
end
def with_preloads
preload(:job_artifacts_archive, :job_artifacts, :tags, project: [:namespace])
end
def clone_accessors
%i[pipeline project ref tag options name
allow_failure stage_idx trigger_request
yaml_variables when environment coverage_regex
description tag_list protected needs_attributes
job_variables_attributes resource_group scheduling_type
ci_stage partition_id id_tokens interruptible execution_config_id].freeze
end
def supported_keyset_orderings
{ id: [:desc] }
end
end
state_machine :status do
event :enqueue do
transition [:created, :skipped, :manual, :scheduled] => :preparing, if: :any_unmet_prerequisites?
end
event :enqueue_scheduled do
transition scheduled: :preparing, if: :any_unmet_prerequisites?
transition scheduled: :pending
end
event :enqueue_preparing do
transition preparing: :pending
end
event :actionize do
transition created: :manual
end
event :schedule do
transition created: :scheduled
end
event :unschedule do
transition scheduled: :manual
end
before_transition on: :enqueue_scheduled do |build|
build.scheduled_at.nil? || build.scheduled_at.past? # If false is returned, it stops the transition
end
before_transition scheduled: any do |build|
build.scheduled_at = nil
end
before_transition created: :scheduled do |build|
build.scheduled_at = build.options_scheduled_at
end
before_transition on: :enqueue_preparing do |build|
!build.any_unmet_prerequisites? # If false is returned, it stops the transition
end
before_transition any => [:pending] do |build|
build.ensure_token
true
end
after_transition created: :scheduled do |build|
build.run_after_commit do
Ci::BuildScheduleWorker.perform_at(build.scheduled_at, build.id)
end
end
after_transition any => [:preparing] do |build|
build.run_after_commit do
Ci::BuildPrepareWorker.perform_async(id)
end
end
# rubocop:disable CodeReuse/ServiceClass
after_transition any => [:pending] do |build, transition|
Ci::UpdateBuildQueueService.new.push(build, transition)
build.run_after_commit do
BuildQueueWorker.perform_async(id)
build.execute_hooks
end
end
after_transition pending: any do |build, transition|
Ci::UpdateBuildQueueService.new.pop(build, transition)
end
after_transition any => [:running] do |build, transition|
Ci::UpdateBuildQueueService.new.track(build, transition)
end
after_transition running: any do |build, transition|
Ci::UpdateBuildQueueService.new.untrack(build, transition)
Ci::BuildRunnerSession.where(build: build).delete_all
end
# rubocop:enable CodeReuse/ServiceClass
#
after_transition pending: :running do |build|
build.ensure_metadata.update_timeout_state
end
after_transition pending: :running do |build|
build.run_after_commit do
build.ensure_persistent_ref
build.execute_hooks
end
end
after_transition any => [:success, :failed, :canceled] do |build|
build.run_after_commit do
build.run_status_commit_hooks!
Ci::BuildFinishedWorker.perform_async(id)
observe_report_types
end
end
after_transition any => [:success] do |build|
build.run_after_commit do
PagesWorker.perform_async(:deploy, id) if build.pages_generator?
end
end
after_transition any => [:failed] do |build|
next unless build.project
build.run_after_commit do
if build.auto_retry_allowed?
begin
# rubocop: disable CodeReuse/ServiceClass -- https://gitlab.com/gitlab-org/gitlab/-/issues/494865
Ci::RetryJobService.new(build.project, build.user).execute(build)
# rubocop: enable CodeReuse/ServiceClass
rescue Gitlab::Access::AccessDeniedError => e
Gitlab::AppLogger.error "Unable to auto-retry job #{build.id}: #{e}"
end
end
end
end
end
def self.build_matchers(project)
unique_params = [
:protected,
Arel.sql("(#{arel_tag_names_array.to_sql})")
]
group(*unique_params).pluck('array_agg(id)', *unique_params).map do |values|
Gitlab::Ci::Matching::BuildMatcher.new({
build_ids: values[0],
protected: values[1],
tag_list: values[2],
project: project
})
end
end
def self.ids_in_merge_request(merge_request_id)
in_merge_request(merge_request_id).pluck(:id)
end
def self.taggings_join_model
::Ci::BuildTag
end
# A Ci::Bridge may transition to `canceling` as a result of strategy: :depend
# but only a Ci::Build will transition to `canceling`` via `.cancel`
def supports_canceling?
cancel_gracefully?
end
def cancel_gracefully?
!!runner_manager&.supports_after_script_on_cancel?
end
def supports_force_cancel?
true
end
def build_matcher
strong_memoize(:build_matcher) do
Gitlab::Ci::Matching::BuildMatcher.new({
protected: protected?,
tag_list: tag_list,
build_ids: [id],
project: project
})
end
end
def auto_retry_allowed?
auto_retry.allowed?
end
def exit_code=(value)
return unless value
ensure_metadata.exit_code = value.to_i.clamp(0, Gitlab::Database::MAX_SMALLINT_VALUE)
end
def auto_retry_expected?
failed? && auto_retry_allowed?
end
def detailed_status(current_user)
Gitlab::Ci::Status::Build::Factory
.new(present, current_user)
.fabricate!
end
def other_scheduled_actions
pipeline.scheduled_actions.reject { |action| action.name == name }
end
def pages_generator?
return false unless Gitlab.config.pages.enabled
return false unless options.present?
return true if options[:pages].is_a?(Hash) || options[:pages] == true
options[:pages] != false && name == 'pages' # Legacy behaviour
end
def pages
return {} unless pages_generator? && publish_path_available?
{ publish: expanded_publish_path }
end
strong_memoize_attr :pages
def runnable?
true
end
def degenerated?
super && execution_config_id.nil?
end
def degenerate!
super do
execution_config&.destroy
end
end
def archived?
degenerated? || super
end
def playable?
action? && !archived? && (manual? || scheduled? || retryable?)
end
def schedulable?
self.when == 'delayed' && options[:start_in].present?
end
def options_scheduled_at
ChronicDuration.parse(options[:start_in])&.seconds&.from_now
end
def action?
ACTIONABLE_WHEN.include?(self.when)
end
def can_auto_cancel_pipeline_on_job_failure?
# A job that doesn't need to be auto-retried can auto-cancel its own pipeline
!auto_retry_expected?
end
# rubocop: disable CodeReuse/ServiceClass
def play(current_user, job_variables_attributes = nil)
Ci::PlayBuildService.new(current_user: current_user, build: self, variables: job_variables_attributes).execute
end
# rubocop: enable CodeReuse/ServiceClass
def cancelable?
(active? || created?) && !canceling?
end
def force_cancelable?
canceling? && supports_force_cancel?
end
def retries_count
pipeline.builds.retried.where(name: name).count
end
override :all_met_to_become_pending?
def all_met_to_become_pending?
super && !any_unmet_prerequisites?
end
def any_unmet_prerequisites?
prerequisites.present?
end
def prerequisites
Gitlab::Ci::Build::Prerequisite::Factory.new(self).unmet
end
def triggered_by?(current_user)
user == current_user
end
##
# All variables, including persisted environment variables.
#
def variables
strong_memoize(:variables) do
Gitlab::Ci::Variables::Collection.new
.concat(base_variables)
.concat(pages_variables)
end
end
def persisted_variables
Gitlab::Ci::Variables::Collection.new.tap do |variables|
break variables unless persisted?
variables
.concat(pipeline.persisted_variables)
.append(key: 'CI_JOB_ID', value: id.to_s)
.append(key: 'CI_JOB_URL', value: Gitlab::Routing.url_helpers.project_job_url(project, self))
.append(key: 'CI_JOB_TOKEN', value: token.to_s, public: false, masked: true)
.append(key: 'CI_JOB_STARTED_AT', value: started_at&.iso8601)
.append(key: 'CI_REGISTRY_USER', value: ::Gitlab::Auth::CI_JOB_USER)
.append(key: 'CI_REGISTRY_PASSWORD', value: token.to_s, public: false, masked: true)
.append(key: 'CI_REPOSITORY_URL', value: repo_url.to_s, public: false)
.concat(deploy_token_variables)
.concat(harbor_variables)
.concat(apple_app_store_variables)
.concat(google_play_variables)
.concat(diffblue_cover_variables)
end
end
def persisted_environment_variables
Gitlab::Ci::Variables::Collection.new.tap do |variables|
break variables unless persisted? && persisted_environment.present?
variables.append(key: 'CI_ENVIRONMENT_SLUG', value: environment_slug)
# Here we're passing unexpanded environment_url for runner to expand,
# and we need to make sure that CI_ENVIRONMENT_NAME and
# CI_ENVIRONMENT_SLUG so on are available for the URL be expanded.
variables.append(key: 'CI_ENVIRONMENT_URL', value: environment_url) if environment_url
end
end
def deploy_token_variables
Gitlab::Ci::Variables::Collection.new.tap do |variables|
break variables unless gitlab_deploy_token
variables.append(key: 'CI_DEPLOY_USER', value: gitlab_deploy_token.username)
variables.append(key: 'CI_DEPLOY_PASSWORD', value: gitlab_deploy_token.token, public: false, masked: true)
end
end
def dependency_proxy_variables
Gitlab::Ci::Variables::Collection.new.tap do |variables|
break variables unless Gitlab.config.dependency_proxy.enabled
variables.append(key: 'CI_DEPENDENCY_PROXY_USER', value: ::Gitlab::Auth::CI_JOB_USER)
variables.append(key: 'CI_DEPENDENCY_PROXY_PASSWORD', value: token.to_s, public: false, masked: true)
end
end
def harbor_variables
return [] unless harbor_integration.try(:activated?)
Gitlab::Ci::Variables::Collection.new(harbor_integration.ci_variables)
end
def apple_app_store_variables
return [] unless apple_app_store_integration.try(:activated?)
Gitlab::Ci::Variables::Collection.new(apple_app_store_integration.ci_variables(protected_ref: pipeline.protected_ref?))
end
def google_play_variables
return [] unless google_play_integration.try(:activated?)
Gitlab::Ci::Variables::Collection.new(google_play_integration.ci_variables(protected_ref: pipeline.protected_ref?))
end
def diffblue_cover_variables
return [] unless diffblue_cover_integration.try(:activated?)
Gitlab::Ci::Variables::Collection.new(diffblue_cover_integration.ci_variables)
end
def pages_variables
::Gitlab::Ci::Variables::Collection.new.tap do |variables|
variables.append(key: 'CI_PAGES_URL', value: project.pages_url(pages))
end
end
def base_variables
::Gitlab::Ci::Variables::Collection.new
.concat(persisted_variables)
.concat(dependency_proxy_variables)
.concat(job_jwt_variables)
.concat(scoped_variables)
.concat(job_variables)
.concat(persisted_environment_variables)
end
strong_memoize_attr :base_variables
def features
{
trace_sections: true,
failure_reasons: self.class.failure_reasons.keys
}
end
def merge_request
strong_memoize(:merge_request) do
pipeline.all_merge_requests.order(iid: :asc).first
end
end
def repo_url
return unless token
auth = "#{::Gitlab::Auth::CI_JOB_USER}:#{token}@"
project.http_url_to_repo.sub(%r{^https?://}) do |prefix|
prefix + auth
end
end
def allow_git_fetch
project.build_allow_git_fetch
end
def update_coverage
coverage = trace.extract_coverage(coverage_regex)
update(coverage: coverage) if coverage.present?
end
def trace
Gitlab::Ci::Trace.new(self)
end
def has_trace?
trace.exist?
end
def has_live_trace?
trace.live?
end
def has_archived_trace?
trace.archived?
end
def artifacts_file
job_artifacts_archive&.file
end
def artifacts_size
job_artifacts_archive&.size
end
def artifacts_metadata
job_artifacts_metadata&.file
end
def artifacts?
!artifacts_expired? && artifacts_file&.exists?
end
def locked_artifacts?
pipeline.artifacts_locked? && artifacts_file&.exists?
end
# This method is similar to #artifacts? but it includes the artifacts
# locking mechanics. A new method was created to prevent breaking existing
# behavior and avoid introducing N+1s.
def available_artifacts?
(!artifacts_expired? || pipeline.artifacts_locked?) && job_artifacts_archive&.exists?
end
def artifacts_metadata?
artifacts? && artifacts_metadata&.exists?
end
def has_job_artifacts?
job_artifacts.any?
end
def has_test_reports?
job_artifacts.of_report_type(:test).exists?
end
def ensure_trace_metadata!
Ci::BuildTraceMetadata.find_or_upsert_for!(id, partition_id)
end
def artifacts_expose_as
options.dig(:artifacts, :expose_as)
end
def artifacts_paths
options.dig(:artifacts, :paths)
end
def needs_touch?
Time.current - updated_at > 15.minutes.to_i
end
def remove_token!
update!(token_encrypted: nil)
end
def has_tags?
tag_list.any?
end
def any_runners_online?
cache_for_online_runners do
project.any_online_runners? { |runner| runner.match_build_if_online?(self) }
end
end
def any_runners_available?
cache_for_available_runners do
project.active_runners.exists?
end
end
def stuck?
pending? && !any_runners_online?
end
def execute_hooks
return unless project
return if user&.blocked?
return unless project.has_active_hooks?(:job_hooks) || project.has_active_integrations?(:job_hooks)
Ci::ExecuteBuildHooksWorker.perform_async(project.id, build_data)
end
def browsable_artifacts?
artifacts_metadata?
end
def artifacts_public?
return true if job_artifacts_archive.nil? # To backward compatibility return true if no artifacts found
job_artifacts_archive.public_access?
end
def artifacts_no_access?
return false if job_artifacts_archive.nil? # To backward compatibility return false if no artifacts found
job_artifacts_archive.none_access?
end
def artifact_access_setting_in_config
artifacts_public = options.dig(:artifacts, :public)
artifacts_access = options.dig(:artifacts, :access)
if !artifacts_public.nil? && !artifacts_access.nil?
raise ArgumentError, 'artifacts:public and artifacts:access are mutually exclusive'
end
return :public if artifacts_public == true || artifacts_access == 'all'
return :private if artifacts_public == false || artifacts_access == 'developer'
return :none if artifacts_access == 'none'
# default behaviour
:public
end
def artifacts_metadata_entry(path, **options)
artifacts_metadata.open do |metadata_stream|
metadata = Gitlab::Ci::Build::Artifacts::Metadata.new(
metadata_stream,
path,
**options)
metadata.to_entry
end
end
def erasable?
complete? && (artifacts? || has_job_artifacts? || has_trace?)
end
def erased?
!erased_at.nil?
end
def artifacts_expired?
artifacts_expire_at&.past?
end
def artifacts_expire_in
artifacts_expire_at - Time.current if artifacts_expire_at
end
def artifacts_expire_in=(value)
self.artifacts_expire_at =
(ChronicDuration.parse(value)&.seconds&.from_now if value)
end
def has_expired_locked_archive_artifacts?
locked_artifacts? &&
artifacts_expire_at&.past?
end
def has_expiring_archive_artifacts?
has_expiring_artifacts? && job_artifacts_archive.present?
end
def self.keep_artifacts!
update_all(artifacts_expire_at: nil)
Ci::JobArtifact.where(job: self.select(:id)).update_all(expire_at: nil)
end
def keep_artifacts!
update(artifacts_expire_at: nil)
job_artifacts.update_all(expire_at: nil)
end
def artifact_for_type(type)
file_types = Ci::JobArtifact.associated_file_types_for(type)
file_types_ids = file_types&.map { |file_type| Ci::JobArtifact.file_types[file_type] }
job_artifacts.find_by(file_type: file_types_ids)
end
def steps
[Gitlab::Ci::Build::Step.from_commands(self),
Gitlab::Ci::Build::Step.from_release(self),
Gitlab::Ci::Build::Step.from_after_script(self)].compact
end
def runtime_hooks
Gitlab::Ci::Build::Hook.from_hooks(self)
end
def image
Gitlab::Ci::Build::Image.from_image(self)
end
def services
Gitlab::Ci::Build::Image.from_services(self)
end
def cache
cache = Array.wrap(options[:cache])
cache.each do |single_cache|
single_cache[:fallback_keys] = [] unless single_cache.key?(:fallback_keys)
end
if project.jobs_cache_index
cache = cache.map do |single_cache|
cache = single_cache.merge(key: "#{single_cache[:key]}-#{project.jobs_cache_index}")
fallback = cache.slice(:fallback_keys).transform_values { |keys| keys.map { |key| "#{key}-#{project.jobs_cache_index}" } }
cache.merge(fallback.compact)
end
end
return cache unless project.ci_separated_caches
cache.map do |entry|
type_suffix = !entry[:unprotect] && pipeline.protected_ref? ? 'protected' : 'non_protected'
cache = entry.merge(key: "#{entry[:key]}-#{type_suffix}")
fallback = cache.slice(:fallback_keys).transform_values { |keys| keys.map { |key| "#{key}-#{type_suffix}" } }
cache.merge(fallback.compact)
end
end
def fallback_cache_keys_defined?
Array.wrap(options[:cache]).any? { |cache| cache[:fallback_keys].present? }
end
def credentials
Gitlab::Ci::Build::Credentials::Factory.new(self).create!
end
def has_valid_build_dependencies?
dependencies.valid?
end
def invalid_dependencies
dependencies.invalid_local
end
def valid_dependency?
return false if artifacts_expired? && !pipeline.artifacts_locked?
return false if erased?
true
end
def runner_required_feature_names
strong_memoize(:runner_required_feature_names) do
RUNNER_FEATURES.select do |feature, method|
method.call(self)
end.keys
end
end
def supported_runner?(features)
runner_required_feature_names.all? do |feature_name|
features&.dig(feature_name)
end
end
def publishes_artifacts_reports?
options&.dig(:artifacts, :reports)&.any?
end
def supports_artifacts_exclude?
options&.dig(:artifacts, :exclude)&.any?
end
def publish_path
return unless options.present?
return options[:publish] unless options[:pages].is_a?(Hash)
options.dig(:pages, :publish) || options[:publish]
end
def publish_path_available?
publish_path.present?
end
def expanded_publish_path
ExpandVariables.expand(publish_path.to_s, -> { base_variables.sort_and_expand_all })
end
def multi_build_steps?
options[:release]&.any?
end
def hide_secrets(data, metrics = ::Gitlab::Ci::Trace::Metrics.new)
return unless trace
data.dup.tap do |trace|
Gitlab::Ci::MaskSecret.mask!(trace, project.runners_token) if project
Gitlab::Ci::MaskSecret.mask!(trace, token) if token
metrics.increment_trace_operation(operation: :mutated) if trace != data
end
end
def serializable_hash(options = {})
super(options).merge(when: read_attribute(:when))
end
def has_terminal?
running? && runner_session_url.present?
end
def collect_test_reports!(test_reports)
each_report(Ci::JobArtifact.file_types_for_report(:test)) do |file_type, blob|
Gitlab::Ci::Parsers.fabricate!(file_type).parse!(blob, test_reports, job: self)
end
test_reports
end
def collect_accessibility_reports!(accessibility_report)
each_report(Ci::JobArtifact.file_types_for_report(:accessibility)) do |file_type, blob|
Gitlab::Ci::Parsers.fabricate!(file_type).parse!(blob, accessibility_report)
end
accessibility_report
end
def collect_codequality_reports!(codequality_report)
each_report(Ci::JobArtifact.file_types_for_report(:codequality)) do |file_type, blob|
Gitlab::Ci::Parsers.fabricate!(file_type).parse!(blob, codequality_report, { project: project, commit_sha: pipeline.sha })
end
codequality_report
end
def collect_terraform_reports!(terraform_reports)
each_report(::Ci::JobArtifact.file_types_for_report(:terraform)) do |file_type, blob, report_artifact|
::Gitlab::Ci::Parsers.fabricate!(file_type).parse!(blob, terraform_reports, artifact: report_artifact)
end
terraform_reports
end
def report_artifacts
job_artifacts.all_reports
end
# Consider this object to have an unknown job problem
def doom!
transaction do
now = Time.current
attributes = {
status: :failed,
failure_reason: :data_integrity_failure,
updated_at: now
}
attributes[:finished_at] = now unless finished_at.present?
update_columns(attributes)
all_queuing_entries.delete_all
all_runtime_metadata.delete_all
end
deployment&.sync_status_with(self)
::Gitlab::Ci::Pipeline::Metrics
.job_failure_reason_counter
.increment(reason: :data_integrity_failure)
Gitlab::AppLogger.info(
message: 'Build doomed',
class: self.class.name,
build_id: id,
pipeline_id: pipeline_id,
project_id: project_id)
end
def degradation_threshold
var = yaml_variables.find { |v| v[:key] == DEGRADATION_THRESHOLD_VARIABLE_NAME }
var[:value]&.to_i if var
end
def remove_pending_state!
pending_state.try(:delete)
end
def run_on_status_commit(&block)
status_commit_hooks.push(block)
end
def max_test_cases_per_report
# NOTE: This is temporary and will be replaced later by a value
# that would come from an actual application limit.
::Gitlab.com? ? 500_000 : 0
end
def debug_mode?
# perform the check on both sides in case the runner version is old
metadata&.debug_trace_enabled? ||
Gitlab::Utils.to_boolean(variables['CI_DEBUG_SERVICES']&.value, default: false) ||
Gitlab::Utils.to_boolean(variables['CI_DEBUG_TRACE']&.value, default: false)
end
def drop_with_exit_code!(failure_reason, exit_code)
failure_reason ||= :unknown_failure
result = drop!(::Gitlab::Ci::Build::Status::Reason.new(self, failure_reason, exit_code))
::Ci::TrackFailedBuildWorker.perform_async(id, exit_code, failure_reason)
result
end
def exit_codes_defined?
options.dig(:allow_failure_criteria, :exit_codes).present? || options.dig(:retry, :exit_codes).present?
end
def create_queuing_entry!
::Ci::PendingBuild.upsert_from_build!(self)
end
##
# We can have only one queuing entry or running build tracking entry,
# because there is a unique index on `build_id` in each table, but we need
# a relation to remove these entries more efficiently in a single statement
# without actually loading data.
#
def all_queuing_entries
::Ci::PendingBuild.where(build_id: id)
end
def all_runtime_metadata
::Ci::RunningBuild.where(build_id: id)
end
def shared_runner_build?
runner&.instance_type?
end
def job_variables_attributes
strong_memoize(:job_variables_attributes) do
job_variables.internal_source.map do |variable|
variable.attributes.except('id', 'job_id', 'encrypted_value', 'encrypted_value_iv').tap do |attrs|
attrs[:value] = variable.value
end
end
end
end
def allowed_to_fail_with_code?(exit_code)
options
.dig(:allow_failure_criteria, :exit_codes)
.to_a
.include?(exit_code)
end
def each_report(report_types)
job_artifacts_for_types(report_types).each do |report_artifact|
next if report_artifact&.artifact_report&.faulty?
report_artifact.each_blob do |blob|
yield report_artifact.file_type, blob, report_artifact
end
end
end
def clone(current_user:, new_job_variables_attributes: [])
new_build = super
if action? && new_job_variables_attributes.any?
new_build.job_variables = []
new_build.job_variables_attributes = new_job_variables_attributes
end
new_build
end
def job_artifact_types
job_artifacts.map(&:file_type)
end
def test_suite_name
if matrix_build?
name
else
group_name
end
end
def time_in_queue_seconds
return if queued_at.nil?
(::Time.current - queued_at).seconds.to_i
end
strong_memoize_attr :time_in_queue_seconds
def source
build_source&.source || pipeline.source
end
strong_memoize_attr :source
# Can be removed in Rails 7.1. Related to: Gitlab.next_rails?
def to_partial_path
'jobs/job'
end
def token
return encoded_jwt if user&.has_composite_identity? || use_jwt_for_ci_cd_job_token?
super
end
protected
def run_status_commit_hooks!
status_commit_hooks.reverse_each do |hook|
instance_eval(&hook)
end
end
private
def use_jwt_for_ci_cd_job_token?
namespace&.root_ancestor&.namespace_settings&.jwt_ci_cd_job_token_enabled?
end
def encoded_jwt
::Ci::JobToken::Jwt.encode(self)
end
strong_memoize_attr :encoded_jwt
def matrix_build?
options.dig(:parallel, :matrix).present?
end
def stick_build_if_status_changed
return unless saved_change_to_status?
return unless running?
self.class.sticking.stick(:build, id)
end
def status_commit_hooks
@status_commit_hooks ||= []
end
def auto_retry
strong_memoize(:auto_retry) do
Gitlab::Ci::Build::AutoRetry.new(self)
end
end
def build_data
strong_memoize(:build_data) do
ActiveRecord::Associations::Preloader.new(records: [self], associations: { runner: :tags }).call
Gitlab::DataBuilder::Build.build(self)
end
end
def job_artifacts_for_types(report_types)
# Use select to leverage cached associations and avoid N+1 queries
job_artifacts.select { |artifact| artifact.file_type.in?(report_types) }
end
def has_expiring_artifacts?
artifacts_expire_at.present? && artifacts_expire_at.future?
end
def job_jwt_variables
id_tokens_variables
end
def id_tokens_variables
Gitlab::Ci::Variables::Collection.new.tap do |variables|
break variables unless id_tokens?
sub_components = project.ci_id_token_sub_claim_components.map(&:to_sym)
id_tokens.each do |var_name, token_data|
token = Gitlab::Ci::JwtV2.for_build(self, aud: expanded_id_token_aud(token_data['aud']),
sub_components: sub_components)
variables.append(key: var_name, value: token, public: false, masked: true)
end
rescue OpenSSL::PKey::RSAError, Gitlab::Ci::Jwt::NoSigningKeyError => e
Gitlab::ErrorTracking.track_exception(e)
end
end
def expanded_id_token_aud(aud)
return unless aud
strong_memoize_with(:expanded_id_token_aud, aud) do
# `aud` can be a string or an array of strings.
if aud.is_a?(Array)
aud.map { |x| ExpandVariables.expand(x, -> { scoped_variables.sort_and_expand_all }) }
else
ExpandVariables.expand(aud, -> { scoped_variables.sort_and_expand_all })
end
end
end
def cache_for_online_runners(&block)
Rails.cache.fetch(
['has-online-runners', id],
expires_in: RUNNERS_STATUS_CACHE_EXPIRATION
) { yield }
end
def cache_for_available_runners(&block)
Rails.cache.fetch(
['has-available-runners', project.id],
expires_in: RUNNERS_STATUS_CACHE_EXPIRATION
) { yield }
end
def observe_report_types
return unless ::Gitlab.com?
report_types = options&.dig(:artifacts, :reports)&.keys || []
report_types.each do |report_type|
next unless Enums::Ci::JobArtifact.report_types.include?(report_type)
::Gitlab::Ci::Artifacts::Metrics
.build_completed_report_type_counter(report_type)
.increment(status: status)
end
end
def track_ci_secrets_management_id_tokens_usage
::Gitlab::UsageDataCounters::HLLRedisCounter.track_event('i_ci_secrets_management_id_tokens_build_created', values: user_id)
Gitlab::Tracking.event(
self.class.to_s,
'create_id_tokens',
namespace: namespace,
user: user,
label: 'redis_hll_counters.ci_secrets_management.i_ci_secrets_management_id_tokens_build_created_monthly',
ultimate_namespace_id: namespace.root_ancestor.id,
context: [Gitlab::Tracking::ServicePingContext.new(
data_source: :redis_hll,
event: 'i_ci_secrets_management_id_tokens_build_created'
).to_context]
)
end
def track_ci_build_created_event
Gitlab::InternalEvents.track_event(
'create_ci_build',
project: project,
user: user,
additional_properties: {
property: name
}
)
end
def partition_id_prefix_in_16_bit_encode
"#{partition_id.to_s(16)}_"
end
def prefix_and_partition_for_token
TOKEN_PREFIX + partition_id_prefix_in_16_bit_encode
end
end
end
Ci::Build.prepend_mod_with('Ci::Build')