diff options
author | GitLab Bot <gitlab-bot@gitlab.com> | 2021-06-16 18:25:58 +0000 |
---|---|---|
committer | GitLab Bot <gitlab-bot@gitlab.com> | 2021-06-16 18:25:58 +0000 |
commit | a5f4bba440d7f9ea47046a0a561d49adf0a1e6d4 (patch) | |
tree | fb69158581673816a8cd895f9d352dcb3c678b1e /spec/lib/bulk_imports/groups/pipelines | |
parent | d16b2e8639e99961de6ddc93909f3bb5c1445ba1 (diff) | |
download | gitlab-ce-a5f4bba440d7f9ea47046a0a561d49adf0a1e6d4.tar.gz |
Add latest changes from gitlab-org/gitlab@14-0-stable-eev14.0.0-rc42
Diffstat (limited to 'spec/lib/bulk_imports/groups/pipelines')
4 files changed, 145 insertions, 150 deletions
diff --git a/spec/lib/bulk_imports/groups/pipelines/boards_pipeline_spec.rb b/spec/lib/bulk_imports/groups/pipelines/boards_pipeline_spec.rb new file mode 100644 index 00000000000..8b2f03ca15f --- /dev/null +++ b/spec/lib/bulk_imports/groups/pipelines/boards_pipeline_spec.rb @@ -0,0 +1,49 @@ +# frozen_string_literal: true + +require 'spec_helper' + +RSpec.describe BulkImports::Groups::Pipelines::BoardsPipeline do + let_it_be(:user) { create(:user) } + let_it_be(:group) { create(:group) } + let_it_be(:bulk_import) { create(:bulk_import, user: user) } + let_it_be(:filepath) { 'spec/fixtures/bulk_imports/gz/boards.ndjson.gz' } + let_it_be(:entity) do + create( + :bulk_import_entity, + group: group, + bulk_import: bulk_import, + source_full_path: 'source/full/path', + destination_name: 'My Destination Group', + destination_namespace: group.full_path + ) + end + + let_it_be(:tracker) { create(:bulk_import_tracker, entity: entity) } + let_it_be(:context) { BulkImports::Pipeline::Context.new(tracker) } + + let(:tmpdir) { Dir.mktmpdir } + + before do + FileUtils.copy_file(filepath, File.join(tmpdir, 'boards.ndjson.gz')) + group.add_owner(user) + end + + subject { described_class.new(context) } + + describe '#run' do + it 'imports group boards into destination group and removes tmpdir' do + allow(Dir).to receive(:mktmpdir).and_return(tmpdir) + allow_next_instance_of(BulkImports::FileDownloadService) do |service| + allow(service).to receive(:execute) + end + + expect { subject.run }.to change(Board, :count).by(1) + + lists = group.boards.find_by(name: 'first board').lists + + expect(lists.count).to eq(3) + expect(lists.first.label.title).to eq('TSL') + expect(lists.second.label.title).to eq('Sosync') + end + end +end diff --git a/spec/lib/bulk_imports/groups/pipelines/entity_finisher_spec.rb b/spec/lib/bulk_imports/groups/pipelines/entity_finisher_spec.rb index 8276349c5f4..b97aeb435b9 100644 --- a/spec/lib/bulk_imports/groups/pipelines/entity_finisher_spec.rb +++ b/spec/lib/bulk_imports/groups/pipelines/entity_finisher_spec.rb @@ -25,13 +25,33 @@ RSpec.describe BulkImports::Groups::Pipelines::EntityFinisher do .to change(entity, :status_name).to(:finished) end - it 'does nothing when the entity is already finished' do - entity = create(:bulk_import_entity, :finished) - pipeline_tracker = create(:bulk_import_tracker, entity: entity) - context = BulkImports::Pipeline::Context.new(pipeline_tracker) - subject = described_class.new(context) + context 'when entity is in a final finished or failed state' do + shared_examples 'performs no state update' do |entity_state| + it 'does nothing' do + entity = create(:bulk_import_entity, entity_state) + pipeline_tracker = create(:bulk_import_tracker, entity: entity) + context = BulkImports::Pipeline::Context.new(pipeline_tracker) + subject = described_class.new(context) - expect { subject.run } - .not_to change(entity, :status_name) + expect { subject.run } + .not_to change(entity, :status_name) + end + end + + include_examples 'performs no state update', :finished + include_examples 'performs no state update', :failed + end + + context 'when all entity trackers failed' do + it 'marks entity as failed' do + entity = create(:bulk_import_entity, :started) + create(:bulk_import_tracker, :failed, entity: entity) + pipeline_tracker = create(:bulk_import_tracker, entity: entity, relation: described_class) + context = BulkImports::Pipeline::Context.new(pipeline_tracker) + + described_class.new(context).run + + expect(entity.reload.failed?).to eq(true) + end end end diff --git a/spec/lib/bulk_imports/groups/pipelines/labels_pipeline_spec.rb b/spec/lib/bulk_imports/groups/pipelines/labels_pipeline_spec.rb index 8af646d1101..6344dae0fb7 100644 --- a/spec/lib/bulk_imports/groups/pipelines/labels_pipeline_spec.rb +++ b/spec/lib/bulk_imports/groups/pipelines/labels_pipeline_spec.rb @@ -5,98 +5,74 @@ require 'spec_helper' RSpec.describe BulkImports::Groups::Pipelines::LabelsPipeline do let_it_be(:user) { create(:user) } let_it_be(:group) { create(:group) } - let_it_be(:timestamp) { Time.new(2020, 01, 01).utc } - + let_it_be(:bulk_import) { create(:bulk_import, user: user) } + let_it_be(:filepath) { 'spec/fixtures/bulk_imports/gz/labels.ndjson.gz' } let_it_be(:entity) do create( :bulk_import_entity, + group: group, + bulk_import: bulk_import, source_full_path: 'source/full/path', destination_name: 'My Destination Group', - destination_namespace: group.full_path, - group: group + destination_namespace: group.full_path ) end let_it_be(:tracker) { create(:bulk_import_tracker, entity: entity) } let_it_be(:context) { BulkImports::Pipeline::Context.new(tracker) } + let(:tmpdir) { Dir.mktmpdir } + + before do + FileUtils.copy_file(filepath, File.join(tmpdir, 'labels.ndjson.gz')) + group.add_owner(user) + end + subject { described_class.new(context) } describe '#run' do - it 'imports a group labels' do - first_page = extracted_data(title: 'label1', has_next_page: true) - last_page = extracted_data(title: 'label2') - - allow_next_instance_of(BulkImports::Common::Extractors::GraphqlExtractor) do |extractor| - allow(extractor) - .to receive(:extract) - .and_return(first_page, last_page) + it 'imports group labels into destination group and removes tmpdir' do + allow(Dir).to receive(:mktmpdir).and_return(tmpdir) + allow_next_instance_of(BulkImports::FileDownloadService) do |service| + allow(service).to receive(:execute) end - expect { subject.run }.to change(Label, :count).by(2) + expect { subject.run }.to change(::GroupLabel, :count).by(1) - label = group.labels.order(:created_at).last + label = group.labels.first - expect(label.title).to eq('label2') - expect(label.description).to eq('desc') - expect(label.color).to eq('#428BCA') - expect(label.created_at).to eq(timestamp) - expect(label.updated_at).to eq(timestamp) + expect(label.title).to eq('Label 1') + expect(label.description).to eq('Label 1') + expect(label.color).to eq('#6699cc') + expect(File.directory?(tmpdir)).to eq(false) end end describe '#load' do - it 'creates the label' do - data = label_data('label') + context 'when label is not persisted' do + it 'saves the label' do + label = build(:group_label, group: group) - expect { subject.load(context, data) }.to change(Label, :count).by(1) + expect(label).to receive(:save!) - label = group.labels.first - - data.each do |key, value| - expect(label[key]).to eq(value) + subject.load(context, label) end end - end - describe 'pipeline parts' do - it { expect(described_class).to include_module(BulkImports::Pipeline) } - it { expect(described_class).to include_module(BulkImports::Pipeline::Runner) } - - it 'has extractors' do - expect(described_class.get_extractor) - .to eq( - klass: BulkImports::Common::Extractors::GraphqlExtractor, - options: { - query: BulkImports::Groups::Graphql::GetLabelsQuery - } - ) - end + context 'when label is persisted' do + it 'does not save label' do + label = create(:group_label, group: group) - it 'has transformers' do - expect(described_class.transformers) - .to contain_exactly( - { klass: BulkImports::Common::Transformers::ProhibitedAttributesTransformer, options: nil } - ) - end - end - - def label_data(title) - { - 'title' => title, - 'description' => 'desc', - 'color' => '#428BCA', - 'created_at' => timestamp.to_s, - 'updated_at' => timestamp.to_s - } - end + expect(label).not_to receive(:save!) - def extracted_data(title:, has_next_page: false) - page_info = { - 'has_next_page' => has_next_page, - 'next_page' => has_next_page ? 'cursor' : nil - } + subject.load(context, label) + end + end - BulkImports::Pipeline::ExtractedData.new(data: [label_data(title)], page_info: page_info) + context 'when label is missing' do + it 'returns' do + expect(subject.load(context, nil)).to be_nil + end + end end end diff --git a/spec/lib/bulk_imports/groups/pipelines/milestones_pipeline_spec.rb b/spec/lib/bulk_imports/groups/pipelines/milestones_pipeline_spec.rb index e5cf75c566b..a8354e62459 100644 --- a/spec/lib/bulk_imports/groups/pipelines/milestones_pipeline_spec.rb +++ b/spec/lib/bulk_imports/groups/pipelines/milestones_pipeline_spec.rb @@ -5,119 +5,69 @@ require 'spec_helper' RSpec.describe BulkImports::Groups::Pipelines::MilestonesPipeline do let_it_be(:user) { create(:user) } let_it_be(:group) { create(:group) } - let_it_be(:timestamp) { Time.new(2020, 01, 01).utc } let_it_be(:bulk_import) { create(:bulk_import, user: user) } - + let_it_be(:filepath) { 'spec/fixtures/bulk_imports/gz/milestones.ndjson.gz' } let_it_be(:entity) do create( :bulk_import_entity, + group: group, bulk_import: bulk_import, source_full_path: 'source/full/path', destination_name: 'My Destination Group', - destination_namespace: group.full_path, - group: group + destination_namespace: group.full_path ) end let_it_be(:tracker) { create(:bulk_import_tracker, entity: entity) } let_it_be(:context) { BulkImports::Pipeline::Context.new(tracker) } - subject { described_class.new(context) } + let(:tmpdir) { Dir.mktmpdir } before do + FileUtils.copy_file(filepath, File.join(tmpdir, 'milestones.ndjson.gz')) group.add_owner(user) end - describe '#run' do - it 'imports group milestones' do - first_page = extracted_data(title: 'milestone1', iid: 1, has_next_page: true) - last_page = extracted_data(title: 'milestone2', iid: 2) + subject { described_class.new(context) } - allow_next_instance_of(BulkImports::Common::Extractors::GraphqlExtractor) do |extractor| - allow(extractor) - .to receive(:extract) - .and_return(first_page, last_page) + describe '#run' do + it 'imports group milestones into destination group and removes tmpdir' do + allow(Dir).to receive(:mktmpdir).and_return(tmpdir) + allow_next_instance_of(BulkImports::FileDownloadService) do |service| + allow(service).to receive(:execute) end - expect { subject.run }.to change(Milestone, :count).by(2) - - expect(group.milestones.pluck(:title)).to contain_exactly('milestone1', 'milestone2') - - milestone = group.milestones.last - - expect(milestone.description).to eq('desc') - expect(milestone.state).to eq('closed') - expect(milestone.start_date.to_s).to eq('2020-10-21') - expect(milestone.due_date.to_s).to eq('2020-10-22') - expect(milestone.created_at).to eq(timestamp) - expect(milestone.updated_at).to eq(timestamp) + expect { subject.run }.to change(Milestone, :count).by(5) + expect(group.milestones.pluck(:title)).to contain_exactly('v4.0', 'v3.0', 'v2.0', 'v1.0', 'v0.0') + expect(File.directory?(tmpdir)).to eq(false) end end describe '#load' do - it 'creates the milestone' do - data = milestone_data('milestone') - - expect { subject.load(context, data) }.to change(Milestone, :count).by(1) - end - - context 'when user is not authorized to create the milestone' do - before do - allow(user).to receive(:can?).with(:admin_milestone, group).and_return(false) - end + context 'when milestone is not persisted' do + it 'saves the milestone' do + milestone = build(:milestone, group: group) - it 'raises NotAllowedError' do - data = extracted_data(title: 'milestone') + expect(milestone).to receive(:save!) - expect { subject.load(context, data) }.to raise_error(::BulkImports::Pipeline::NotAllowedError) + subject.load(context, milestone) end end - end - describe 'pipeline parts' do - it { expect(described_class).to include_module(BulkImports::Pipeline) } - it { expect(described_class).to include_module(BulkImports::Pipeline::Runner) } + context 'when milestone is persisted' do + it 'does not save milestone' do + milestone = create(:milestone, group: group) - it 'has extractors' do - expect(described_class.get_extractor) - .to eq( - klass: BulkImports::Common::Extractors::GraphqlExtractor, - options: { - query: BulkImports::Groups::Graphql::GetMilestonesQuery - } - ) - end + expect(milestone).not_to receive(:save!) - it 'has transformers' do - expect(described_class.transformers) - .to contain_exactly( - { klass: BulkImports::Common::Transformers::ProhibitedAttributesTransformer, options: nil } - ) + subject.load(context, milestone) + end end - end - def milestone_data(title, iid: 1) - { - 'title' => title, - 'description' => 'desc', - 'iid' => iid, - 'state' => 'closed', - 'start_date' => '2020-10-21', - 'due_date' => '2020-10-22', - 'created_at' => timestamp.to_s, - 'updated_at' => timestamp.to_s - } - end - - def extracted_data(title:, iid: 1, has_next_page: false) - page_info = { - 'has_next_page' => has_next_page, - 'next_page' => has_next_page ? 'cursor' : nil - } - - BulkImports::Pipeline::ExtractedData.new( - data: milestone_data(title, iid: iid), - page_info: page_info - ) + context 'when milestone is missing' do + it 'returns' do + expect(subject.load(context, nil)).to be_nil + end + end end end |