# frozen_string_literal: true require 'spec_helper' RSpec.describe Ci::CreateJobArtifactsService do let_it_be(:project) { create(:project) } let(:service) { described_class.new(job) } let(:job) { create(:ci_build, project: project) } let(:artifacts_sha256) { '0' * 64 } let(:metadata_file) { nil } let(:artifacts_file) do file_to_upload('spec/fixtures/ci_build_artifacts.zip', sha256: artifacts_sha256) end let(:params) do { 'artifact_type' => 'archive', 'artifact_format' => 'zip' }.with_indifferent_access end def file_to_upload(path, params = {}) upload = Tempfile.new('upload') FileUtils.copy(path, upload.path) UploadedFile.new(upload.path, **params) end def unique_metrics_report_uploaders Gitlab::UsageDataCounters::HLLRedisCounter.unique_events( event_names: described_class::METRICS_REPORT_UPLOAD_EVENT_NAME, start_date: 2.weeks.ago, end_date: 2.weeks.from_now ) end describe '#execute' do subject { service.execute(artifacts_file, params, metadata_file: metadata_file) } context 'when artifacts file is uploaded' do it 'saves artifact for the given type' do expect { subject }.to change { Ci::JobArtifact.count }.by(1) new_artifact = job.job_artifacts.last expect(new_artifact.project).to eq(job.project) expect(new_artifact.file).to be_present expect(new_artifact.file_type).to eq(params['artifact_type']) expect(new_artifact.file_format).to eq(params['artifact_format']) expect(new_artifact.file_sha256).to eq(artifacts_sha256) end it 'does not track the job user_id' do subject expect(unique_metrics_report_uploaders).to eq(0) end context 'when metadata file is also uploaded' do let(:metadata_file) do file_to_upload('spec/fixtures/ci_build_artifacts_metadata.gz', sha256: artifacts_sha256) end before do stub_application_setting(default_artifacts_expire_in: '1 day') end it 'saves metadata artifact' do expect { subject }.to change { Ci::JobArtifact.count }.by(2) new_artifact = job.job_artifacts.last expect(new_artifact.project).to eq(job.project) expect(new_artifact.file).to be_present expect(new_artifact.file_type).to eq('metadata') expect(new_artifact.file_format).to eq('gzip') expect(new_artifact.file_sha256).to eq(artifacts_sha256) end it 'sets expiration date according to application settings' do expected_expire_at = 1.day.from_now expect(subject).to match(a_hash_including(status: :success)) archive_artifact, metadata_artifact = job.job_artifacts.last(2) expect(job.artifacts_expire_at).to be_within(1.minute).of(expected_expire_at) expect(archive_artifact.expire_at).to be_within(1.minute).of(expected_expire_at) expect(metadata_artifact.expire_at).to be_within(1.minute).of(expected_expire_at) end context 'when expire_in params is set to a specific value' do before do params.merge!('expire_in' => '2 hours') end it 'sets expiration date according to the parameter' do expected_expire_at = 2.hours.from_now expect(subject).to match(a_hash_including(status: :success)) archive_artifact, metadata_artifact = job.job_artifacts.last(2) expect(job.artifacts_expire_at).to be_within(1.minute).of(expected_expire_at) expect(archive_artifact.expire_at).to be_within(1.minute).of(expected_expire_at) expect(metadata_artifact.expire_at).to be_within(1.minute).of(expected_expire_at) end end context 'when expire_in params is set to `never`' do before do params.merge!('expire_in' => 'never') end it 'sets expiration date according to the parameter' do expected_expire_at = nil expect(subject).to be_truthy archive_artifact, metadata_artifact = job.job_artifacts.last(2) expect(job.artifacts_expire_at).to eq(expected_expire_at) expect(archive_artifact.expire_at).to eq(expected_expire_at) expect(metadata_artifact.expire_at).to eq(expected_expire_at) end end end end context 'when artifacts file already exists' do let!(:existing_artifact) do create(:ci_job_artifact, :archive, file_sha256: existing_sha256, job: job) end context 'when sha256 of uploading artifact is the same of the existing one' do let(:existing_sha256) { artifacts_sha256 } it 'ignores the changes' do expect { subject }.not_to change { Ci::JobArtifact.count } expect(subject).to match(a_hash_including(status: :success)) end end context 'when sha256 of uploading artifact is different than the existing one' do let(:existing_sha256) { '1' * 64 } it 'returns error status' do expect(Gitlab::ErrorTracking).to receive(:track_exception).and_call_original expect { subject }.not_to change { Ci::JobArtifact.count } expect(subject).to match( a_hash_including(http_status: :bad_request, message: 'another artifact of the same type already exists', status: :error)) end end end context 'when artifact type is dotenv' do let(:artifacts_file) do file_to_upload('spec/fixtures/build.env.gz', sha256: artifacts_sha256) end let(:params) do { 'artifact_type' => 'dotenv', 'artifact_format' => 'gzip' }.with_indifferent_access end it 'calls parse service' do expect_any_instance_of(Ci::ParseDotenvArtifactService) do |service| expect(service).to receive(:execute).once.and_call_original end expect(subject[:status]).to eq(:success) expect(job.job_variables.as_json).to contain_exactly( hash_including('key' => 'KEY1', 'value' => 'VAR1', 'source' => 'dotenv'), hash_including('key' => 'KEY2', 'value' => 'VAR2', 'source' => 'dotenv')) end context 'when ci_synchronous_artifact_parsing feature flag is disabled' do before do stub_feature_flags(ci_synchronous_artifact_parsing: false) end it 'does not call parse service' do expect(Ci::ParseDotenvArtifactService).not_to receive(:new) expect(subject[:status]).to eq(:success) end end end context 'when artifact_type is metrics' do before do allow(job).to receive(:user_id).and_return(123) end let(:params) { { 'artifact_type' => 'metrics', 'artifact_format' => 'gzip' }.with_indifferent_access } it 'tracks the job user_id' do subject expect(unique_metrics_report_uploaders).to eq(1) end end context 'when artifact type is cluster_applications' do let(:artifacts_file) do file_to_upload('spec/fixtures/helm/helm_list_v2_prometheus_missing.json.gz', sha256: artifacts_sha256) end let(:params) do { 'artifact_type' => 'cluster_applications', 'artifact_format' => 'gzip' }.with_indifferent_access end it 'calls cluster applications parse service' do expect_next_instance_of(Clusters::ParseClusterApplicationsArtifactService) do |service| expect(service).to receive(:execute).once.and_call_original end subject end context 'when there is a deployment cluster' do let(:user) { project.owner } before do job.update!(user: user) end it 'calls cluster applications parse service with job and job user', :aggregate_failures do expect(Clusters::ParseClusterApplicationsArtifactService).to receive(:new).with(job, user).and_call_original subject end end context 'when ci_synchronous_artifact_parsing feature flag is disabled' do before do stub_feature_flags(ci_synchronous_artifact_parsing: false) end it 'does not call parse service' do expect(Clusters::ParseClusterApplicationsArtifactService).not_to receive(:new) expect(subject[:status]).to eq(:success) end end end shared_examples 'rescues object storage error' do |klass, message, expected_message| it "handles #{klass}" do allow_next_instance_of(JobArtifactUploader) do |uploader| allow(uploader).to receive(:store!).and_raise(klass, message) end expect(Gitlab::ErrorTracking) .to receive(:track_exception) .and_call_original expect(subject).to match( a_hash_including( http_status: :service_unavailable, message: expected_message || message, status: :error)) end end it_behaves_like 'rescues object storage error', Errno::EIO, 'some/path', 'Input/output error - some/path' it_behaves_like 'rescues object storage error', Google::Apis::ServerError, 'Server error' it_behaves_like 'rescues object storage error', Signet::RemoteServerError, 'The service is currently unavailable' end end