diff options
author | GitLab Bot <gitlab-bot@gitlab.com> | 2020-04-08 03:09:31 +0000 |
---|---|---|
committer | GitLab Bot <gitlab-bot@gitlab.com> | 2020-04-08 03:09:31 +0000 |
commit | e2ee1eec50aa8df8543d7ecc585ec0ba5ee544ac (patch) | |
tree | 7998650d27ada12ee7d06a21cbb3b5e89f298378 /spec/lib | |
parent | 060c842402c00f830a810702600cbe39dfa6cf62 (diff) | |
download | gitlab-ce-e2ee1eec50aa8df8543d7ecc585ec0ba5ee544ac.tar.gz |
Add latest changes from gitlab-org/gitlab@master
Diffstat (limited to 'spec/lib')
-rw-r--r-- | spec/lib/gitlab/graphql/pagination/externally_paginated_array_connection_spec.rb (renamed from spec/lib/gitlab/graphql/connections/externally_paginated_array_connection_spec.rb) | 18 | ||||
-rw-r--r-- | spec/lib/gitlab/graphql/pagination/filterable_array_connection_spec.rb (renamed from spec/lib/gitlab/graphql/connections/filterable_array_connection_spec.rb) | 12 | ||||
-rw-r--r-- | spec/lib/gitlab/graphql/pagination/keyset/conditions/not_null_condition_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/conditions/not_null_condition_spec.rb) | 4 | ||||
-rw-r--r-- | spec/lib/gitlab/graphql/pagination/keyset/conditions/null_condition_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/conditions/null_condition_spec.rb) | 4 | ||||
-rw-r--r-- | spec/lib/gitlab/graphql/pagination/keyset/connection_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/connection_spec.rb) | 19 | ||||
-rw-r--r-- | spec/lib/gitlab/graphql/pagination/keyset/order_info_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/order_info_spec.rb) | 2 | ||||
-rw-r--r-- | spec/lib/gitlab/graphql/pagination/keyset/query_builder_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/query_builder_spec.rb) | 4 | ||||
-rw-r--r-- | spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb | 70 | ||||
-rw-r--r-- | spec/lib/gitlab/import_export/json/ndjson_reader_spec.rb | 113 | ||||
-rw-r--r-- | spec/lib/gitlab/import_export/project/tree_restorer_spec.rb | 1390 | ||||
-rw-r--r-- | spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb | 41 |
11 files changed, 948 insertions, 729 deletions
diff --git a/spec/lib/gitlab/graphql/connections/externally_paginated_array_connection_spec.rb b/spec/lib/gitlab/graphql/pagination/externally_paginated_array_connection_spec.rb index 83c94ed6260..85a5b1dacc7 100644 --- a/spec/lib/gitlab/graphql/connections/externally_paginated_array_connection_spec.rb +++ b/spec/lib/gitlab/graphql/pagination/externally_paginated_array_connection_spec.rb @@ -2,7 +2,7 @@ require 'spec_helper' -describe Gitlab::Graphql::Connections::ExternallyPaginatedArrayConnection do +describe Gitlab::Graphql::Pagination::ExternallyPaginatedArrayConnection do let(:prev_cursor) { 1 } let(:next_cursor) { 6 } let(:values) { [2, 3, 4, 5] } @@ -10,21 +10,13 @@ describe Gitlab::Graphql::Connections::ExternallyPaginatedArrayConnection do let(:arguments) { {} } subject(:connection) do - described_class.new(all_nodes, arguments) + described_class.new(all_nodes, { max_page_size: values.size }.merge(arguments)) end - describe '#sliced_nodes' do - let(:sliced_nodes) { connection.sliced_nodes } + describe '#nodes' do + let(:paged_nodes) { connection.nodes } - it 'returns all the nodes' do - expect(connection.sliced_nodes).to eq(values) - end - end - - describe '#paged_nodes' do - let(:paged_nodes) { connection.send(:paged_nodes) } - - it_behaves_like "connection with paged nodes" do + it_behaves_like 'connection with paged nodes' do let(:paged_nodes_size) { values.size } end end diff --git a/spec/lib/gitlab/graphql/connections/filterable_array_connection_spec.rb b/spec/lib/gitlab/graphql/pagination/filterable_array_connection_spec.rb index b2f0862be62..c82e3ad3019 100644 --- a/spec/lib/gitlab/graphql/connections/filterable_array_connection_spec.rb +++ b/spec/lib/gitlab/graphql/pagination/filterable_array_connection_spec.rb @@ -2,19 +2,19 @@ require 'spec_helper' -describe Gitlab::Graphql::Connections::FilterableArrayConnection do +describe Gitlab::Graphql::Pagination::FilterableArrayConnection do let(:callback) { proc { |nodes| nodes } } let(:all_nodes) { Gitlab::Graphql::FilterableArray.new(callback, 1, 2, 3, 4, 5) } let(:arguments) { {} } subject(:connection) do - described_class.new(all_nodes, arguments, max_page_size: 3) + described_class.new(all_nodes, { max_page_size: 3 }.merge(arguments)) end - describe '#paged_nodes' do - let(:paged_nodes) { subject.paged_nodes } + describe '#nodes' do + let(:paged_nodes) { subject.nodes } - it_behaves_like "connection with paged nodes" do + it_behaves_like 'connection with paged nodes' do let(:paged_nodes_size) { 3 } end @@ -22,7 +22,7 @@ describe Gitlab::Graphql::Connections::FilterableArrayConnection do let(:callback) { proc { |nodes| nodes[1..-1] } } it 'does not return filtered elements' do - expect(subject.paged_nodes).to contain_exactly(all_nodes[1], all_nodes[2]) + expect(subject.nodes).to contain_exactly(all_nodes[1], all_nodes[2]) end end end diff --git a/spec/lib/gitlab/graphql/connections/keyset/conditions/not_null_condition_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/conditions/not_null_condition_spec.rb index 26fc5344871..33dc6bdea97 100644 --- a/spec/lib/gitlab/graphql/connections/keyset/conditions/not_null_condition_spec.rb +++ b/spec/lib/gitlab/graphql/pagination/keyset/conditions/not_null_condition_spec.rb @@ -2,7 +2,7 @@ require 'spec_helper' -describe Gitlab::Graphql::Connections::Keyset::Conditions::NotNullCondition do +describe Gitlab::Graphql::Pagination::Keyset::Conditions::NotNullCondition do describe '#build' do let(:operators) { ['>', '>'] } let(:before_or_after) { :after } @@ -75,7 +75,7 @@ describe Gitlab::Graphql::Connections::Keyset::Conditions::NotNullCondition do context 'when ordering by LOWER' do let(:arel_table) { Project.arel_table } let(:relation) { Project.order(arel_table['name'].lower.asc).order(:id) } - let(:order_list) { Gitlab::Graphql::Connections::Keyset::OrderInfo.build_order_list(relation) } + let(:order_list) { Gitlab::Graphql::Pagination::Keyset::OrderInfo.build_order_list(relation) } let(:values) { ['Test', 500] } context 'when :after' do diff --git a/spec/lib/gitlab/graphql/connections/keyset/conditions/null_condition_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/conditions/null_condition_spec.rb index be0a21b2438..1c74f2fb0ab 100644 --- a/spec/lib/gitlab/graphql/connections/keyset/conditions/null_condition_spec.rb +++ b/spec/lib/gitlab/graphql/pagination/keyset/conditions/null_condition_spec.rb @@ -2,7 +2,7 @@ require 'spec_helper' -describe Gitlab::Graphql::Connections::Keyset::Conditions::NullCondition do +describe Gitlab::Graphql::Pagination::Keyset::Conditions::NullCondition do describe '#build' do let(:values) { [nil, 500] } let(:operators) { [nil, '>'] } @@ -58,7 +58,7 @@ describe Gitlab::Graphql::Connections::Keyset::Conditions::NullCondition do context 'when ordering by LOWER' do let(:arel_table) { Project.arel_table } let(:relation) { Project.order(arel_table['name'].lower.asc).order(:id) } - let(:order_list) { Gitlab::Graphql::Connections::Keyset::OrderInfo.build_order_list(relation) } + let(:order_list) { Gitlab::Graphql::Pagination::Keyset::OrderInfo.build_order_list(relation) } context 'when :after' do it 'generates sql' do diff --git a/spec/lib/gitlab/graphql/connections/keyset/connection_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/connection_spec.rb index c193ab2b50f..fdacecbaca6 100644 --- a/spec/lib/gitlab/graphql/connections/keyset/connection_spec.rb +++ b/spec/lib/gitlab/graphql/pagination/keyset/connection_spec.rb @@ -2,25 +2,28 @@ require 'spec_helper' -describe Gitlab::Graphql::Connections::Keyset::Connection do +describe Gitlab::Graphql::Pagination::Keyset::Connection do let(:nodes) { Project.all.order(id: :asc) } let(:arguments) { {} } + let(:query_type) { GraphQL::ObjectType.new } + let(:schema) { GraphQL::Schema.define(query: query_type, mutation: nil)} + let(:context) { GraphQL::Query::Context.new(query: OpenStruct.new(schema: schema), values: nil, object: nil) } subject(:connection) do - described_class.new(nodes, arguments, max_page_size: 3) + described_class.new(nodes, { context: context, max_page_size: 3 }.merge(arguments)) end def encoded_cursor(node) - described_class.new(nodes, {}).cursor_from_node(node) + described_class.new(nodes, { context: context }).cursor_for(node) end def decoded_cursor(cursor) JSON.parse(Base64Bp.urlsafe_decode64(cursor)) end - describe '#cursor_from_nodes' do + describe '#cursor_for' do let(:project) { create(:project) } - let(:cursor) { connection.cursor_from_node(project) } + let(:cursor) { connection.cursor_for(project) } it 'returns an encoded ID' do expect(decoded_cursor(cursor)).to eq('id' => project.id.to_s) @@ -264,11 +267,11 @@ describe Gitlab::Graphql::Connections::Keyset::Connection do end end - describe '#paged_nodes' do + describe '#nodes' do let_it_be(:all_nodes) { create_list(:project, 5) } - let(:paged_nodes) { subject.paged_nodes } + let(:paged_nodes) { subject.nodes } - it_behaves_like "connection with paged nodes" do + it_behaves_like 'connection with paged nodes' do let(:paged_nodes_size) { 3 } end diff --git a/spec/lib/gitlab/graphql/connections/keyset/order_info_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/order_info_spec.rb index eb823fc0122..cb5656d7c00 100644 --- a/spec/lib/gitlab/graphql/connections/keyset/order_info_spec.rb +++ b/spec/lib/gitlab/graphql/pagination/keyset/order_info_spec.rb @@ -2,7 +2,7 @@ require 'spec_helper' -describe Gitlab::Graphql::Connections::Keyset::OrderInfo do +describe Gitlab::Graphql::Pagination::Keyset::OrderInfo do describe '#build_order_list' do let(:order_list) { described_class.build_order_list(relation) } diff --git a/spec/lib/gitlab/graphql/connections/keyset/query_builder_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/query_builder_spec.rb index b46ce4bf023..7fd5b889cf9 100644 --- a/spec/lib/gitlab/graphql/connections/keyset/query_builder_spec.rb +++ b/spec/lib/gitlab/graphql/pagination/keyset/query_builder_spec.rb @@ -2,7 +2,7 @@ require 'spec_helper' -describe Gitlab::Graphql::Connections::Keyset::QueryBuilder do +describe Gitlab::Graphql::Pagination::Keyset::QueryBuilder do context 'when number of ordering fields is 0' do it 'raises an error' do expect { described_class.new(Issue.arel_table, [], {}, :after) } @@ -12,7 +12,7 @@ describe Gitlab::Graphql::Connections::Keyset::QueryBuilder do describe '#conditions' do let(:relation) { Issue.order(relative_position: :desc).order(:id) } - let(:order_list) { Gitlab::Graphql::Connections::Keyset::OrderInfo.build_order_list(relation) } + let(:order_list) { Gitlab::Graphql::Pagination::Keyset::OrderInfo.build_order_list(relation) } let(:arel_table) { Issue.arel_table } let(:builder) { described_class.new(arel_table, order_list, decoded_cursor, before_or_after) } let(:before_or_after) { :after } diff --git a/spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb b/spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb index 297a5946703..3e9bd3fe741 100644 --- a/spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb +++ b/spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb @@ -15,7 +15,6 @@ RSpec.shared_examples 'import/export json legacy reader' do subject { legacy_reader.consume_attributes("project") } context 'no excluded attributes' do - let(:excluded_attributes) { [] } let(:relation_names) { [] } it 'returns the whole tree from parsed JSON' do @@ -42,60 +41,53 @@ RSpec.shared_examples 'import/export json legacy reader' do describe '#consume_relation' do context 'when valid path is passed' do - let(:key) { 'description' } + let(:key) { 'labels' } - context 'block not given' do - it 'returns value of the key' do - expect(legacy_reader).to receive(:relations).and_return({ key => 'test value' }) - expect(legacy_reader.consume_relation("project", key)).to eq('test value') - end - end + subject { legacy_reader.consume_relation("project", key) } - context 'key has been consumed' do - before do - legacy_reader.consume_relation("project", key) + context 'key has not been consumed' do + it 'returns an Enumerator' do + expect(subject).to be_an_instance_of(Enumerator) end - it 'does not yield' do - expect do |blk| - legacy_reader.consume_relation("project", key, &blk) - end.not_to yield_control - end - end + context 'value is nil' do + before do + expect(legacy_reader).to receive(:relations).and_return({ key => nil }) + end - context 'value is nil' do - before do - expect(legacy_reader).to receive(:relations).and_return({ key => nil }) + it 'yields nothing to the Enumerator' do + expect(subject.to_a).to eq([]) + end end - it 'does not yield' do - expect do |blk| - legacy_reader.consume_relation("project", key, &blk) - end.not_to yield_control - end - end + context 'value is an array' do + before do + expect(legacy_reader).to receive(:relations).and_return({ key => %w[label1 label2] }) + end - context 'value is not array' do - before do - expect(legacy_reader).to receive(:relations).and_return({ key => 'value' }) + it 'yields every relation value to the Enumerator' do + expect(subject.to_a).to eq([['label1', 0], ['label2', 1]]) + end end - it 'yield the value with index 0' do - expect do |blk| - legacy_reader.consume_relation("project", key, &blk) - end.to yield_with_args('value', 0) + context 'value is not array' do + before do + expect(legacy_reader).to receive(:relations).and_return({ key => 'non-array value' }) + end + + it 'yields the value with index 0 to the Enumerator' do + expect(subject.to_a).to eq([['non-array value', 0]]) + end end end - context 'value is an array' do + context 'key has been consumed' do before do - expect(legacy_reader).to receive(:relations).and_return({ key => %w[item1 item2 item3] }) + legacy_reader.consume_relation("project", key).first end - it 'yield each array element with index' do - expect do |blk| - legacy_reader.consume_relation("project", key, &blk) - end.to yield_successive_args(['item1', 0], ['item2', 1], ['item3', 2]) + it 'yields nothing to the Enumerator' do + expect(subject.to_a).to eq([]) end end end diff --git a/spec/lib/gitlab/import_export/json/ndjson_reader_spec.rb b/spec/lib/gitlab/import_export/json/ndjson_reader_spec.rb new file mode 100644 index 00000000000..40b784fdb87 --- /dev/null +++ b/spec/lib/gitlab/import_export/json/ndjson_reader_spec.rb @@ -0,0 +1,113 @@ +# frozen_string_literal: true + +require 'spec_helper' + +describe Gitlab::ImportExport::JSON::NdjsonReader do + include ImportExport::CommonUtil + + let(:fixture) { 'spec/fixtures/lib/gitlab/import_export/light/tree' } + let(:root_tree) { JSON.parse(File.read(File.join(fixture, 'project.json'))) } + let(:ndjson_reader) { described_class.new(dir_path) } + let(:importable_path) { 'project' } + + before :all do + extract_archive('spec/fixtures/lib/gitlab/import_export/light', 'tree.tar.gz') + end + + after :all do + cleanup_artifacts_from_extract_archive('light') + end + + describe '#exist?' do + subject { ndjson_reader.exist? } + + context 'given valid dir_path' do + let(:dir_path) { fixture } + + it { is_expected.to be true } + end + + context 'given invalid dir_path' do + let(:dir_path) { 'invalid-dir-path' } + + it { is_expected.to be false } + end + end + + describe '#legacy?' do + let(:dir_path) { fixture } + + subject { ndjson_reader.legacy? } + + it { is_expected.to be false } + end + + describe '#consume_attributes' do + let(:dir_path) { fixture } + + subject { ndjson_reader.consume_attributes(importable_path) } + + it 'returns the whole root tree from parsed JSON' do + expect(subject).to eq(root_tree) + end + end + + describe '#consume_relation' do + let(:dir_path) { fixture } + + subject { ndjson_reader.consume_relation(importable_path, key) } + + context 'given any key' do + let(:key) { 'any-key' } + + it 'returns an Enumerator' do + expect(subject).to be_an_instance_of(Enumerator) + end + end + + context 'key has been consumed' do + let(:key) { 'issues' } + + before do + ndjson_reader.consume_relation(importable_path, key).first + end + + it 'yields nothing to the Enumerator' do + expect(subject.to_a).to eq([]) + end + end + + context 'key has not been consumed' do + context 'relation file does not exist' do + let(:key) { 'non-exist-relation-file-name' } + + before do + relation_file_path = File.join(dir_path, importable_path, "#{key}.ndjson") + expect(File).to receive(:exist?).with(relation_file_path).and_return(false) + end + + it 'yields nothing to the Enumerator' do + expect(subject.to_a).to eq([]) + end + end + + context 'relation file is empty' do + let(:key) { 'empty' } + + it 'yields nothing to the Enumerator' do + expect(subject.to_a).to eq([]) + end + end + + context 'relation file contains multiple lines' do + let(:key) { 'custom_attributes' } + let(:attr_1) { JSON.parse('{"id":201,"project_id":5,"created_at":"2016-06-14T15:01:51.315Z","updated_at":"2016-06-14T15:01:51.315Z","key":"color","value":"red"}') } + let(:attr_2) { JSON.parse('{"id":202,"project_id":5,"created_at":"2016-06-14T15:01:51.315Z","updated_at":"2016-06-14T15:01:51.315Z","key":"size","value":"small"}') } + + it 'yields every relation value to the Enumerator' do + expect(subject.to_a).to eq([[attr_1, 0], [attr_2, 1]]) + end + end + end + end +end diff --git a/spec/lib/gitlab/import_export/project/tree_restorer_spec.rb b/spec/lib/gitlab/import_export/project/tree_restorer_spec.rb index 247d455fecc..96aed774cfc 100644 --- a/spec/lib/gitlab/import_export/project/tree_restorer_spec.rb +++ b/spec/lib/gitlab/import_export/project/tree_restorer_spec.rb @@ -11,76 +11,83 @@ describe Gitlab::ImportExport::Project::TreeRestorer do let(:shared) { project.import_export_shared } - describe 'restore project tree' do - before_all do - # Using an admin for import, so we can check assignment of existing members - @user = create(:admin) - @existing_members = [ - create(:user, email: 'bernard_willms@gitlabexample.com'), - create(:user, email: 'saul_will@gitlabexample.com') - ] + RSpec.shared_examples 'project tree restorer work properly' do |reader| + describe 'restore project tree' do + before_all do + # Using an admin for import, so we can check assignment of existing members + @user = create(:admin) + @existing_members = [ + create(:user, email: 'bernard_willms@gitlabexample.com'), + create(:user, email: 'saul_will@gitlabexample.com') + ] - RSpec::Mocks.with_temporary_scope do - @project = create(:project, :builds_enabled, :issues_disabled, name: 'project', path: 'project') - @shared = @project.import_export_shared + RSpec::Mocks.with_temporary_scope do + @project = create(:project, :builds_enabled, :issues_disabled, name: 'project', path: 'project') + @shared = @project.import_export_shared - setup_import_export_config('complex') + setup_import_export_config('complex') + setup_reader(reader) - allow_any_instance_of(Repository).to receive(:fetch_source_branch!).and_return(true) - allow_any_instance_of(Gitlab::Git::Repository).to receive(:branch_exists?).and_return(false) + allow_any_instance_of(Repository).to receive(:fetch_source_branch!).and_return(true) + allow_any_instance_of(Gitlab::Git::Repository).to receive(:branch_exists?).and_return(false) - expect_any_instance_of(Gitlab::Git::Repository).to receive(:create_branch).with('feature', 'DCBA') - allow_any_instance_of(Gitlab::Git::Repository).to receive(:create_branch) + expect_any_instance_of(Gitlab::Git::Repository).to receive(:create_branch).with('feature', 'DCBA') + allow_any_instance_of(Gitlab::Git::Repository).to receive(:create_branch) - project_tree_restorer = described_class.new(user: @user, shared: @shared, project: @project) + project_tree_restorer = described_class.new(user: @user, shared: @shared, project: @project) - @restored_project_json = project_tree_restorer.restore + @restored_project_json = project_tree_restorer.restore + end end - end - context 'JSON' do - it 'restores models based on JSON' do - expect(@restored_project_json).to be_truthy + after(:context) do + cleanup_artifacts_from_extract_archive('complex') end - it 'restore correct project features' do - project = Project.find_by_path('project') + context 'JSON' do + it 'restores models based on JSON' do + expect(@restored_project_json).to be_truthy + end - expect(project.project_feature.issues_access_level).to eq(ProjectFeature::PRIVATE) - expect(project.project_feature.builds_access_level).to eq(ProjectFeature::PRIVATE) - expect(project.project_feature.snippets_access_level).to eq(ProjectFeature::PRIVATE) - expect(project.project_feature.wiki_access_level).to eq(ProjectFeature::PRIVATE) - expect(project.project_feature.merge_requests_access_level).to eq(ProjectFeature::PRIVATE) - end + it 'restore correct project features' do + project = Project.find_by_path('project') - it 'has the project description' do - expect(Project.find_by_path('project').description).to eq('Nisi et repellendus ut enim quo accusamus vel magnam.') - end + expect(project.project_feature.issues_access_level).to eq(ProjectFeature::PRIVATE) + expect(project.project_feature.builds_access_level).to eq(ProjectFeature::PRIVATE) + expect(project.project_feature.snippets_access_level).to eq(ProjectFeature::PRIVATE) + expect(project.project_feature.wiki_access_level).to eq(ProjectFeature::PRIVATE) + expect(project.project_feature.merge_requests_access_level).to eq(ProjectFeature::PRIVATE) + end - it 'has the same label associated to two issues' do - expect(ProjectLabel.find_by_title('test2').issues.count).to eq(2) - end + it 'has the project description' do + expect(Project.find_by_path('project').description).to eq('Nisi et repellendus ut enim quo accusamus vel magnam.') + end - it 'has milestones associated to two separate issues' do - expect(Milestone.find_by_description('test milestone').issues.count).to eq(2) - end + it 'has the same label associated to two issues' do + expect(ProjectLabel.find_by_title('test2').issues.count).to eq(2) + end + + it 'has milestones associated to two separate issues' do + expect(Milestone.find_by_description('test milestone').issues.count).to eq(2) + end - context 'when importing a project with cached_markdown_version and note_html' do - context 'for an Issue' do - it 'does not import note_html' do - note_content = 'Quo reprehenderit aliquam qui dicta impedit cupiditate eligendi' - issue_note = Issue.find_by(description: 'Aliquam enim illo et possimus.').notes.select { |n| n.note.match(/#{note_content}/)}.first + context 'when importing a project with cached_markdown_version and note_html' do + context 'for an Issue' do + it 'does not import note_html' do + note_content = 'Quo reprehenderit aliquam qui dicta impedit cupiditate eligendi' + issue_note = Issue.find_by(description: 'Aliquam enim illo et possimus.').notes.select { |n| n.note.match(/#{note_content}/)}.first - expect(issue_note.note_html).to match(/#{note_content}/) + expect(issue_note.note_html).to match(/#{note_content}/) + end end - end - context 'for a Merge Request' do - it 'does not import note_html' do - note_content = 'Sit voluptatibus eveniet architecto quidem' - merge_request_note = match_mr1_note(note_content) + context 'for a Merge Request' do + it 'does not import note_html' do + note_content = 'Sit voluptatibus eveniet architecto quidem' + merge_request_note = match_mr1_note(note_content) - expect(merge_request_note.note_html).to match(/#{note_content}/) + expect(merge_request_note.note_html).to match(/#{note_content}/) + end end context 'merge request system note metadata' do @@ -103,33 +110,32 @@ describe Gitlab::ImportExport::Project::TreeRestorer do end end end - end - it 'creates a valid pipeline note' do - expect(Ci::Pipeline.find_by_sha('sha-notes').notes).not_to be_empty - end + it 'creates a valid pipeline note' do + expect(Ci::Pipeline.find_by_sha('sha-notes').notes).not_to be_empty + end - it 'pipeline has the correct user ID' do - expect(Ci::Pipeline.find_by_sha('sha-notes').user_id).to eq(@user.id) - end + it 'pipeline has the correct user ID' do + expect(Ci::Pipeline.find_by_sha('sha-notes').user_id).to eq(@user.id) + end - it 'restores pipelines with missing ref' do - expect(Ci::Pipeline.where(ref: nil)).not_to be_empty - end + it 'restores pipelines with missing ref' do + expect(Ci::Pipeline.where(ref: nil)).not_to be_empty + end - it 'restores pipeline for merge request' do - pipeline = Ci::Pipeline.find_by_sha('048721d90c449b244b7b4c53a9186b04330174ec') + it 'restores pipeline for merge request' do + pipeline = Ci::Pipeline.find_by_sha('048721d90c449b244b7b4c53a9186b04330174ec') - expect(pipeline).to be_valid - expect(pipeline.tag).to be_falsey - expect(pipeline.source).to eq('merge_request_event') - expect(pipeline.merge_request.id).to be > 0 - expect(pipeline.merge_request.target_branch).to eq('feature') - expect(pipeline.merge_request.source_branch).to eq('feature_conflict') - end + expect(pipeline).to be_valid + expect(pipeline.tag).to be_falsey + expect(pipeline.source).to eq('merge_request_event') + expect(pipeline.merge_request.id).to be > 0 + expect(pipeline.merge_request.target_branch).to eq('feature') + expect(pipeline.merge_request.source_branch).to eq('feature_conflict') + end - it 'restores pipelines based on ascending id order' do - expected_ordered_shas = %w[ + it 'restores pipelines based on ascending id order' do + expected_ordered_shas = %w[ 2ea1f3dec713d940208fb5ce4a38765ecb5d3f73 ce84140e8b878ce6e7c4d298c7202ff38170e3ac 048721d90c449b244b7b4c53a9186b04330174ec @@ -137,732 +143,749 @@ describe Gitlab::ImportExport::Project::TreeRestorer do 5f923865dde3436854e9ceb9cdb7815618d4e849 d2d430676773caa88cdaf7c55944073b2fd5561a 2ea1f3dec713d940208fb5ce4a38765ecb5d3f73 - ] + ] - project = Project.find_by_path('project') + project = Project.find_by_path('project') - project.ci_pipelines.order(:id).each_with_index do |pipeline, i| - expect(pipeline['sha']).to eq expected_ordered_shas[i] + project.ci_pipelines.order(:id).each_with_index do |pipeline, i| + expect(pipeline['sha']).to eq expected_ordered_shas[i] + end end - end - it 'preserves updated_at on issues' do - issue = Issue.find_by(description: 'Aliquam enim illo et possimus.') + it 'preserves updated_at on issues' do + issue = Issue.find_by(description: 'Aliquam enim illo et possimus.') - expect(issue.reload.updated_at.to_s).to eq('2016-06-14 15:02:47 UTC') - end + expect(issue.reload.updated_at.to_s).to eq('2016-06-14 15:02:47 UTC') + end - it 'has multiple issue assignees' do - expect(Issue.find_by(title: 'Voluptatem').assignees).to contain_exactly(@user, *@existing_members) - expect(Issue.find_by(title: 'Issue without assignees').assignees).to be_empty - end + it 'has multiple issue assignees' do + expect(Issue.find_by(title: 'Voluptatem').assignees).to contain_exactly(@user, *@existing_members) + expect(Issue.find_by(title: 'Issue without assignees').assignees).to be_empty + end - it 'restores timelogs for issues' do - timelog = Issue.find_by(title: 'issue_with_timelogs').timelogs.last + it 'restores timelogs for issues' do + timelog = Issue.find_by(title: 'issue_with_timelogs').timelogs.last - aggregate_failures do - expect(timelog.time_spent).to eq(72000) - expect(timelog.spent_at).to eq("2019-12-27T00:00:00.000Z") + aggregate_failures do + expect(timelog.time_spent).to eq(72000) + expect(timelog.spent_at).to eq("2019-12-27T00:00:00.000Z") + end end - end - it 'contains the merge access levels on a protected branch' do - expect(ProtectedBranch.first.merge_access_levels).not_to be_empty - end + it 'contains the merge access levels on a protected branch' do + expect(ProtectedBranch.first.merge_access_levels).not_to be_empty + end - it 'contains the push access levels on a protected branch' do - expect(ProtectedBranch.first.push_access_levels).not_to be_empty - end + it 'contains the push access levels on a protected branch' do + expect(ProtectedBranch.first.push_access_levels).not_to be_empty + end - it 'contains the create access levels on a protected tag' do - expect(ProtectedTag.first.create_access_levels).not_to be_empty - end + it 'contains the create access levels on a protected tag' do + expect(ProtectedTag.first.create_access_levels).not_to be_empty + end - it 'restores issue resource label events' do - expect(Issue.find_by(title: 'Voluptatem').resource_label_events).not_to be_empty - end + it 'restores issue resource label events' do + expect(Issue.find_by(title: 'Voluptatem').resource_label_events).not_to be_empty + end - it 'restores merge requests resource label events' do - expect(MergeRequest.find_by(title: 'MR1').resource_label_events).not_to be_empty - end + it 'restores merge requests resource label events' do + expect(MergeRequest.find_by(title: 'MR1').resource_label_events).not_to be_empty + end - it 'restores suggestion' do - note = Note.find_by("note LIKE 'Saepe asperiores exercitationem non dignissimos laborum reiciendis et ipsum%'") + it 'restores suggestion' do + note = Note.find_by("note LIKE 'Saepe asperiores exercitationem non dignissimos laborum reiciendis et ipsum%'") - expect(note.suggestions.count).to eq(1) - expect(note.suggestions.first.from_content).to eq("Original line\n") - end + expect(note.suggestions.count).to eq(1) + expect(note.suggestions.first.from_content).to eq("Original line\n") + end - context 'event at forth level of the tree' do - let(:event) { Event.find_by(action: 6) } + context 'event at forth level of the tree' do + let(:event) { Event.find_by(action: 6) } - it 'restores the event' do - expect(event).not_to be_nil - end + it 'restores the event' do + expect(event).not_to be_nil + end - it 'has the action' do - expect(event.action).not_to be_nil - end + it 'has the action' do + expect(event.action).not_to be_nil + end - it 'event belongs to note, belongs to merge request, belongs to a project' do - expect(event.note.noteable.project).not_to be_nil + it 'event belongs to note, belongs to merge request, belongs to a project' do + expect(event.note.noteable.project).not_to be_nil + end end - end - it 'has the correct data for merge request diff files' do - expect(MergeRequestDiffFile.where.not(diff: nil).count).to eq(55) - end + it 'has the correct data for merge request diff files' do + expect(MergeRequestDiffFile.where.not(diff: nil).count).to eq(55) + end - it 'has the correct data for merge request diff commits' do - expect(MergeRequestDiffCommit.count).to eq(77) - end + it 'has the correct data for merge request diff commits' do + expect(MergeRequestDiffCommit.count).to eq(77) + end - it 'has the correct data for merge request latest_merge_request_diff' do - MergeRequest.find_each do |merge_request| - expect(merge_request.latest_merge_request_diff_id).to eq(merge_request.merge_request_diffs.maximum(:id)) + it 'has the correct data for merge request latest_merge_request_diff' do + MergeRequest.find_each do |merge_request| + expect(merge_request.latest_merge_request_diff_id).to eq(merge_request.merge_request_diffs.maximum(:id)) + end end - end - it 'has labels associated to label links, associated to issues' do - expect(Label.first.label_links.first.target).not_to be_nil - end + it 'has labels associated to label links, associated to issues' do + expect(Label.first.label_links.first.target).not_to be_nil + end - it 'has project labels' do - expect(ProjectLabel.count).to eq(3) - end + it 'has project labels' do + expect(ProjectLabel.count).to eq(3) + end - it 'has no group labels' do - expect(GroupLabel.count).to eq(0) - end + it 'has no group labels' do + expect(GroupLabel.count).to eq(0) + end - it 'has issue boards' do - expect(Project.find_by_path('project').boards.count).to eq(1) - end + it 'has issue boards' do + expect(Project.find_by_path('project').boards.count).to eq(1) + end - it 'has lists associated with the issue board' do - expect(Project.find_by_path('project').boards.find_by_name('TestBoardABC').lists.count).to eq(3) - end + it 'has lists associated with the issue board' do + expect(Project.find_by_path('project').boards.find_by_name('TestBoardABC').lists.count).to eq(3) + end - it 'has a project feature' do - expect(@project.project_feature).not_to be_nil - end + it 'has a project feature' do + expect(@project.project_feature).not_to be_nil + end - it 'has custom attributes' do - expect(@project.custom_attributes.count).to eq(2) - end + it 'has custom attributes' do + expect(@project.custom_attributes.count).to eq(2) + end - it 'has badges' do - expect(@project.project_badges.count).to eq(2) - end + it 'has badges' do + expect(@project.project_badges.count).to eq(2) + end - it 'has snippets' do - expect(@project.snippets.count).to eq(1) - end + it 'has snippets' do + expect(@project.snippets.count).to eq(1) + end - it 'has award emoji for a snippet' do - award_emoji = @project.snippets.first.award_emoji + it 'has award emoji for a snippet' do + award_emoji = @project.snippets.first.award_emoji - expect(award_emoji.map(&:name)).to contain_exactly('thumbsup', 'coffee') - end + expect(award_emoji.map(&:name)).to contain_exactly('thumbsup', 'coffee') + end - it 'snippet has notes' do - expect(@project.snippets.first.notes.count).to eq(1) - end + it 'snippet has notes' do + expect(@project.snippets.first.notes.count).to eq(1) + end - it 'snippet has award emojis on notes' do - award_emoji = @project.snippets.first.notes.first.award_emoji.first + it 'snippet has award emojis on notes' do + award_emoji = @project.snippets.first.notes.first.award_emoji.first - expect(award_emoji.name).to eq('thumbsup') - end + expect(award_emoji.name).to eq('thumbsup') + end - it 'restores `ci_cd_settings` : `group_runners_enabled` setting' do - expect(@project.ci_cd_settings.group_runners_enabled?).to eq(false) - end + it 'restores `ci_cd_settings` : `group_runners_enabled` setting' do + expect(@project.ci_cd_settings.group_runners_enabled?).to eq(false) + end - it 'restores `auto_devops`' do - expect(@project.auto_devops_enabled?).to eq(true) - expect(@project.auto_devops.deploy_strategy).to eq('continuous') - end + it 'restores `auto_devops`' do + expect(@project.auto_devops_enabled?).to eq(true) + expect(@project.auto_devops.deploy_strategy).to eq('continuous') + end - it 'restores the correct service' do - expect(CustomIssueTrackerService.first).not_to be_nil - end + it 'restores the correct service' do + expect(CustomIssueTrackerService.first).not_to be_nil + end - it 'restores zoom meetings' do - meetings = @project.issues.first.zoom_meetings + it 'restores zoom meetings' do + meetings = @project.issues.first.zoom_meetings - expect(meetings.count).to eq(1) - expect(meetings.first.url).to eq('https://zoom.us/j/123456789') - end + expect(meetings.count).to eq(1) + expect(meetings.first.url).to eq('https://zoom.us/j/123456789') + end - it 'restores sentry issues' do - sentry_issue = @project.issues.first.sentry_issue + it 'restores sentry issues' do + sentry_issue = @project.issues.first.sentry_issue - expect(sentry_issue.sentry_issue_identifier).to eq(1234567891) - end + expect(sentry_issue.sentry_issue_identifier).to eq(1234567891) + end - it 'has award emoji for an issue' do - award_emoji = @project.issues.first.award_emoji.first + it 'has award emoji for an issue' do + award_emoji = @project.issues.first.award_emoji.first - expect(award_emoji.name).to eq('musical_keyboard') - end + expect(award_emoji.name).to eq('musical_keyboard') + end - it 'has award emoji for a note in an issue' do - award_emoji = @project.issues.first.notes.first.award_emoji.first + it 'has award emoji for a note in an issue' do + award_emoji = @project.issues.first.notes.first.award_emoji.first - expect(award_emoji.name).to eq('clapper') - end + expect(award_emoji.name).to eq('clapper') + end - it 'restores container_expiration_policy' do - policy = Project.find_by_path('project').container_expiration_policy + it 'restores container_expiration_policy' do + policy = Project.find_by_path('project').container_expiration_policy - aggregate_failures do - expect(policy).to be_an_instance_of(ContainerExpirationPolicy) - expect(policy).to be_persisted - expect(policy.cadence).to eq('3month') + aggregate_failures do + expect(policy).to be_an_instance_of(ContainerExpirationPolicy) + expect(policy).to be_persisted + expect(policy.cadence).to eq('3month') + end end - end - it 'restores error_tracking_setting' do - setting = @project.error_tracking_setting + it 'restores error_tracking_setting' do + setting = @project.error_tracking_setting - aggregate_failures do - expect(setting.api_url).to eq("https://gitlab.example.com/api/0/projects/sentry-org/sentry-project") - expect(setting.project_name).to eq("Sentry Project") - expect(setting.organization_name).to eq("Sentry Org") + aggregate_failures do + expect(setting.api_url).to eq("https://gitlab.example.com/api/0/projects/sentry-org/sentry-project") + expect(setting.project_name).to eq("Sentry Project") + expect(setting.organization_name).to eq("Sentry Org") + end end - end - it 'restores external pull requests' do - external_pr = @project.external_pull_requests.last + it 'restores external pull requests' do + external_pr = @project.external_pull_requests.last - aggregate_failures do - expect(external_pr.pull_request_iid).to eq(4) - expect(external_pr.source_branch).to eq("feature") - expect(external_pr.target_branch).to eq("master") - expect(external_pr.status).to eq("open") + aggregate_failures do + expect(external_pr.pull_request_iid).to eq(4) + expect(external_pr.source_branch).to eq("feature") + expect(external_pr.target_branch).to eq("master") + expect(external_pr.status).to eq("open") + end end - end - it 'restores pipeline schedules' do - pipeline_schedule = @project.pipeline_schedules.last + it 'restores pipeline schedules' do + pipeline_schedule = @project.pipeline_schedules.last - aggregate_failures do - expect(pipeline_schedule.description).to eq('Schedule Description') - expect(pipeline_schedule.ref).to eq('master') - expect(pipeline_schedule.cron).to eq('0 4 * * 0') - expect(pipeline_schedule.cron_timezone).to eq('UTC') - expect(pipeline_schedule.active).to eq(true) + aggregate_failures do + expect(pipeline_schedule.description).to eq('Schedule Description') + expect(pipeline_schedule.ref).to eq('master') + expect(pipeline_schedule.cron).to eq('0 4 * * 0') + expect(pipeline_schedule.cron_timezone).to eq('UTC') + expect(pipeline_schedule.active).to eq(true) + end end - end - it 'restores releases with links' do - release = @project.releases.last - link = release.links.last + it 'restores releases with links' do + release = @project.releases.last + link = release.links.last - aggregate_failures do - expect(release.tag).to eq('release-1.1') - expect(release.description).to eq('Some release notes') - expect(release.name).to eq('release-1.1') - expect(release.sha).to eq('901de3a8bd5573f4a049b1457d28bc1592ba6bf9') - expect(release.released_at).to eq('2019-12-26T10:17:14.615Z') + aggregate_failures do + expect(release.tag).to eq('release-1.1') + expect(release.description).to eq('Some release notes') + expect(release.name).to eq('release-1.1') + expect(release.sha).to eq('901de3a8bd5573f4a049b1457d28bc1592ba6bf9') + expect(release.released_at).to eq('2019-12-26T10:17:14.615Z') - expect(link.url).to eq('http://localhost/namespace6/project6/-/jobs/140463678/artifacts/download') - expect(link.name).to eq('release-1.1.dmg') + expect(link.url).to eq('http://localhost/namespace6/project6/-/jobs/140463678/artifacts/download') + expect(link.name).to eq('release-1.1.dmg') + end end - end - context 'Merge requests' do - it 'always has the new project as a target' do - expect(MergeRequest.find_by_title('MR1').target_project).to eq(@project) - end + context 'Merge requests' do + it 'always has the new project as a target' do + expect(MergeRequest.find_by_title('MR1').target_project).to eq(@project) + end - it 'has the same source project as originally if source/target are the same' do - expect(MergeRequest.find_by_title('MR1').source_project).to eq(@project) - end + it 'has the same source project as originally if source/target are the same' do + expect(MergeRequest.find_by_title('MR1').source_project).to eq(@project) + end - it 'has the new project as target if source/target differ' do - expect(MergeRequest.find_by_title('MR2').target_project).to eq(@project) - end + it 'has the new project as target if source/target differ' do + expect(MergeRequest.find_by_title('MR2').target_project).to eq(@project) + end - it 'has no source if source/target differ' do - expect(MergeRequest.find_by_title('MR2').source_project_id).to be_nil - end + it 'has no source if source/target differ' do + expect(MergeRequest.find_by_title('MR2').source_project_id).to be_nil + end - it 'has award emoji' do - award_emoji = MergeRequest.find_by_title('MR1').award_emoji + it 'has award emoji' do + award_emoji = MergeRequest.find_by_title('MR1').award_emoji - expect(award_emoji.map(&:name)).to contain_exactly('thumbsup', 'drum') - end + expect(award_emoji.map(&:name)).to contain_exactly('thumbsup', 'drum') + end - context 'notes' do - it 'has award emoji' do - merge_request_note = match_mr1_note('Sit voluptatibus eveniet architecto quidem') - award_emoji = merge_request_note.award_emoji.first + context 'notes' do + it 'has award emoji' do + merge_request_note = match_mr1_note('Sit voluptatibus eveniet architecto quidem') + award_emoji = merge_request_note.award_emoji.first - expect(award_emoji.name).to eq('tada') + expect(award_emoji.name).to eq('tada') + end end end - end - context 'tokens are regenerated' do - it 'has new CI trigger tokens' do - expect(Ci::Trigger.where(token: %w[cdbfasdf44a5958c83654733449e585 33a66349b5ad01fc00174af87804e40])) - .to be_empty - end + context 'tokens are regenerated' do + it 'has new CI trigger tokens' do + expect(Ci::Trigger.where(token: %w[cdbfasdf44a5958c83654733449e585 33a66349b5ad01fc00174af87804e40])) + .to be_empty + end - it 'has a new CI build token' do - expect(Ci::Build.where(token: 'abcd')).to be_empty + it 'has a new CI build token' do + expect(Ci::Build.where(token: 'abcd')).to be_empty + end end - end - context 'has restored the correct number of records' do - it 'has the correct number of merge requests' do - expect(@project.merge_requests.size).to eq(9) - end + context 'has restored the correct number of records' do + it 'has the correct number of merge requests' do + expect(@project.merge_requests.size).to eq(9) + end - it 'only restores valid triggers' do - expect(@project.triggers.size).to eq(1) - end + it 'only restores valid triggers' do + expect(@project.triggers.size).to eq(1) + end - it 'has the correct number of pipelines and statuses' do - expect(@project.ci_pipelines.size).to eq(7) + it 'has the correct number of pipelines and statuses' do + expect(@project.ci_pipelines.size).to eq(7) - @project.ci_pipelines.order(:id).zip([2, 0, 2, 2, 2, 2, 0]) - .each do |(pipeline, expected_status_size)| - expect(pipeline.statuses.size).to eq(expected_status_size) + @project.ci_pipelines.order(:id).zip([2, 0, 2, 2, 2, 2, 0]) + .each do |(pipeline, expected_status_size)| + expect(pipeline.statuses.size).to eq(expected_status_size) + end end end - end - context 'when restoring hierarchy of pipeline, stages and jobs' do - it 'restores pipelines' do - expect(Ci::Pipeline.all.count).to be 7 - end + context 'when restoring hierarchy of pipeline, stages and jobs' do + it 'restores pipelines' do + expect(Ci::Pipeline.all.count).to be 7 + end - it 'restores pipeline stages' do - expect(Ci::Stage.all.count).to be 6 - end + it 'restores pipeline stages' do + expect(Ci::Stage.all.count).to be 6 + end - it 'correctly restores association between stage and a pipeline' do - expect(Ci::Stage.all).to all(have_attributes(pipeline_id: a_value > 0)) - end + it 'correctly restores association between stage and a pipeline' do + expect(Ci::Stage.all).to all(have_attributes(pipeline_id: a_value > 0)) + end - it 'restores statuses' do - expect(CommitStatus.all.count).to be 10 - end + it 'restores statuses' do + expect(CommitStatus.all.count).to be 10 + end - it 'correctly restores association between a stage and a job' do - expect(CommitStatus.all).to all(have_attributes(stage_id: a_value > 0)) - end + it 'correctly restores association between a stage and a job' do + expect(CommitStatus.all).to all(have_attributes(stage_id: a_value > 0)) + end - it 'correctly restores association between a pipeline and a job' do - expect(CommitStatus.all).to all(have_attributes(pipeline_id: a_value > 0)) - end + it 'correctly restores association between a pipeline and a job' do + expect(CommitStatus.all).to all(have_attributes(pipeline_id: a_value > 0)) + end - it 'restores a Hash for CommitStatus options' do - expect(CommitStatus.all.map(&:options).compact).to all(be_a(Hash)) - end + it 'restores a Hash for CommitStatus options' do + expect(CommitStatus.all.map(&:options).compact).to all(be_a(Hash)) + end - it 'restores external pull request for the restored pipeline' do - pipeline_with_external_pr = @project.ci_pipelines.find_by(source: 'external_pull_request_event') + it 'restores external pull request for the restored pipeline' do + pipeline_with_external_pr = @project.ci_pipelines.find_by(source: 'external_pull_request_event') - expect(pipeline_with_external_pr.external_pull_request).to be_persisted - end + expect(pipeline_with_external_pr.external_pull_request).to be_persisted + end - it 'has no import failures' do - expect(@project.import_failures.size).to eq 0 + it 'has no import failures' do + expect(@project.import_failures.size).to eq 0 + end end end end - end - shared_examples 'restores group correctly' do |**results| - it 'has group label' do - expect(project.group.labels.size).to eq(results.fetch(:labels, 0)) - expect(project.group.labels.where(type: "GroupLabel").where.not(project_id: nil).count).to eq(0) - end + shared_examples 'restores group correctly' do |**results| + it 'has group label' do + expect(project.group.labels.size).to eq(results.fetch(:labels, 0)) + expect(project.group.labels.where(type: "GroupLabel").where.not(project_id: nil).count).to eq(0) + end - it 'has group milestone' do - expect(project.group.milestones.size).to eq(results.fetch(:milestones, 0)) - end + it 'has group milestone' do + expect(project.group.milestones.size).to eq(results.fetch(:milestones, 0)) + end - it 'has the correct visibility level' do - # INTERNAL in the `project.json`, group's is PRIVATE - expect(project.visibility_level).to eq(Gitlab::VisibilityLevel::PRIVATE) + it 'has the correct visibility level' do + # INTERNAL in the `project.json`, group's is PRIVATE + expect(project.visibility_level).to eq(Gitlab::VisibilityLevel::PRIVATE) + end end - end - context 'project.json file access check' do - let(:user) { create(:user) } - let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') } - let(:project_tree_restorer) do - described_class.new(user: user, shared: shared, project: project) - end - let(:restored_project_json) { project_tree_restorer.restore } + context 'project.json file access check' do + let(:user) { create(:user) } + let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') } + let(:project_tree_restorer) do + described_class.new(user: user, shared: shared, project: project) + end + let(:restored_project_json) { project_tree_restorer.restore } - it 'does not read a symlink' do - Dir.mktmpdir do |tmpdir| - setup_symlink(tmpdir, 'project.json') - allow(shared).to receive(:export_path).and_call_original + it 'does not read a symlink' do + Dir.mktmpdir do |tmpdir| + setup_symlink(tmpdir, 'project.json') + allow(shared).to receive(:export_path).and_call_original - expect(project_tree_restorer.restore).to eq(false) - expect(shared.errors).to include('Incorrect JSON format') + expect(project_tree_restorer.restore).to eq(false) + expect(shared.errors).to include('invalid import format') + end end end - end - context 'Light JSON' do - let(:user) { create(:user) } - let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') } - let(:project_tree_restorer) { described_class.new(user: user, shared: shared, project: project) } - let(:restored_project_json) { project_tree_restorer.restore } + context 'Light JSON' do + let(:user) { create(:user) } + let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') } + let(:project_tree_restorer) { described_class.new(user: user, shared: shared, project: project) } + let(:restored_project_json) { project_tree_restorer.restore } - context 'with a simple project' do - before do - setup_import_export_config('light') - expect(restored_project_json).to eq(true) - end + context 'with a simple project' do + before do + setup_import_export_config('light') + setup_reader(reader) + + expect(restored_project_json).to eq(true) + end + + after do + cleanup_artifacts_from_extract_archive('light') + end + + it 'issue system note metadata restored successfully' do + note_content = 'created merge request !1 to address this issue' + note = project.issues.first.notes.select { |n| n.note.match(/#{note_content}/)}.first + + expect(note.noteable_type).to eq('Issue') + expect(note.system).to eq(true) + expect(note.system_note_metadata.action).to eq('merge') + expect(note.system_note_metadata.commit_count).to be_nil + end + + context 'when there is an existing build with build token' do + before do + create(:ci_build, token: 'abcd') + end - it_behaves_like 'restores project successfully', - issues: 1, - labels: 2, - label_with_priorities: 'A project label', - milestones: 1, - first_issue_labels: 1, - services: 1 - - it 'issue system note metadata restored successfully' do - note_content = 'created merge request !1 to address this issue' - note = project.issues.first.notes.select { |n| n.note.match(/#{note_content}/)}.first - - expect(note.noteable_type).to eq('Issue') - expect(note.system).to eq(true) - expect(note.system_note_metadata.action).to eq('merge') - expect(note.system_note_metadata.commit_count).to be_nil + it_behaves_like 'restores project successfully', + issues: 1, + labels: 2, + label_with_priorities: 'A project label', + milestones: 1, + first_issue_labels: 1, + services: 1 + end + + context 'when there is an existing build with build token' do + before do + create(:ci_build, token: 'abcd') + end + + it_behaves_like 'restores project successfully', + issues: 1, + labels: 2, + label_with_priorities: 'A project label', + milestones: 1, + first_issue_labels: 1 + end end - context 'when there is an existing build with build token' do + context 'multiple pipelines reference the same external pull request' do before do - create(:ci_build, token: 'abcd') + setup_import_export_config('multi_pipeline_ref_one_external_pr') + setup_reader(reader) + + expect(restored_project_json).to eq(true) + end + + after do + cleanup_artifacts_from_extract_archive('multi_pipeline_ref_one_external_pr') end it_behaves_like 'restores project successfully', - issues: 1, - labels: 2, - label_with_priorities: 'A project label', - milestones: 1, - first_issue_labels: 1 + issues: 0, + labels: 0, + milestones: 0, + ci_pipelines: 2, + external_pull_requests: 1, + import_failures: 0 + + it 'restores external pull request for the restored pipelines' do + external_pr = project.external_pull_requests.first + + project.ci_pipelines.each do |pipeline_with_external_pr| + expect(pipeline_with_external_pr.external_pull_request).to be_persisted + expect(pipeline_with_external_pr.external_pull_request).to eq(external_pr) + end + end end - end - context 'multiple pipelines reference the same external pull request' do - before do - setup_import_export_config('multi_pipeline_ref_one_external_pr') - expect(restored_project_json).to eq(true) - end + context 'when post import action throw non-retriable exception' do + let(:exception) { StandardError.new('post_import_error') } + + before do + setup_import_export_config('light') + setup_reader(reader) - it_behaves_like 'restores project successfully', - issues: 0, - labels: 0, - milestones: 0, - ci_pipelines: 2, - external_pull_requests: 1, - import_failures: 0 + expect(project) + .to receive(:merge_requests) + .and_raise(exception) + end - it 'restores external pull request for the restored pipelines' do - external_pr = project.external_pull_requests.first + after do + cleanup_artifacts_from_extract_archive('light') + end - project.ci_pipelines.each do |pipeline_with_external_pr| - expect(pipeline_with_external_pr.external_pull_request).to be_persisted - expect(pipeline_with_external_pr.external_pull_request).to eq(external_pr) + it 'report post import error' do + expect(restored_project_json).to eq(false) + expect(shared.errors).to include('post_import_error') end end - end - context 'when post import action throw non-retriable exception' do - let(:exception) { StandardError.new('post_import_error') } + context 'when post import action throw retriable exception one time' do + let(:exception) { GRPC::DeadlineExceeded.new } - before do - setup_import_export_config('light') - expect(project) - .to receive(:merge_requests) - .and_raise(exception) - end + before do + setup_import_export_config('light') + setup_reader(reader) - it 'report post import error' do - expect(restored_project_json).to eq(false) - expect(shared.errors).to include('post_import_error') - end - end + expect(project) + .to receive(:merge_requests) + .and_raise(exception) + expect(project) + .to receive(:merge_requests) + .and_call_original + expect(restored_project_json).to eq(true) + end - context 'when post import action throw retriable exception one time' do - let(:exception) { GRPC::DeadlineExceeded.new } + after do + cleanup_artifacts_from_extract_archive('light') + end - before do - setup_import_export_config('light') - expect(project) - .to receive(:merge_requests) - .and_raise(exception) - expect(project) - .to receive(:merge_requests) - .and_call_original - expect(restored_project_json).to eq(true) - end + it_behaves_like 'restores project successfully', + issues: 1, + labels: 2, + label_with_priorities: 'A project label', + milestones: 1, + first_issue_labels: 1, + services: 1, + import_failures: 1 - it_behaves_like 'restores project successfully', - issues: 1, - labels: 2, - label_with_priorities: 'A project label', - milestones: 1, - first_issue_labels: 1, - services: 1, - import_failures: 1 - - it 'records the failures in the database' do - import_failure = ImportFailure.last - - expect(import_failure.project_id).to eq(project.id) - expect(import_failure.relation_key).to be_nil - expect(import_failure.relation_index).to be_nil - expect(import_failure.exception_class).to eq('GRPC::DeadlineExceeded') - expect(import_failure.exception_message).to be_present - expect(import_failure.correlation_id_value).not_to be_empty - expect(import_failure.created_at).to be_present - end - end + it 'records the failures in the database' do + import_failure = ImportFailure.last - context 'when the project has overridden params in import data' do - before do - setup_import_export_config('light') + expect(import_failure.project_id).to eq(project.id) + expect(import_failure.relation_key).to be_nil + expect(import_failure.relation_index).to be_nil + expect(import_failure.exception_class).to eq('GRPC::DeadlineExceeded') + expect(import_failure.exception_message).to be_present + expect(import_failure.correlation_id_value).not_to be_empty + expect(import_failure.created_at).to be_present + end end - it 'handles string versions of visibility_level' do - # Project needs to be in a group for visibility level comparison - # to happen - group = create(:group) - project.group = group + context 'when the project has overridden params in import data' do + before do + setup_import_export_config('light') + setup_reader(reader) + end - project.create_import_data(data: { override_params: { visibility_level: Gitlab::VisibilityLevel::INTERNAL.to_s } }) + after do + cleanup_artifacts_from_extract_archive('light') + end - expect(restored_project_json).to eq(true) - expect(project.visibility_level).to eq(Gitlab::VisibilityLevel::INTERNAL) - end + it 'handles string versions of visibility_level' do + # Project needs to be in a group for visibility level comparison + # to happen + group = create(:group) + project.group = group - it 'overwrites the params stored in the JSON' do - project.create_import_data(data: { override_params: { description: "Overridden" } }) + project.create_import_data(data: { override_params: { visibility_level: Gitlab::VisibilityLevel::INTERNAL.to_s } }) - expect(restored_project_json).to eq(true) - expect(project.description).to eq("Overridden") - end + expect(restored_project_json).to eq(true) + expect(project.visibility_level).to eq(Gitlab::VisibilityLevel::INTERNAL) + end - it 'does not allow setting params that are excluded from import_export settings' do - project.create_import_data(data: { override_params: { lfs_enabled: true } }) + it 'overwrites the params stored in the JSON' do + project.create_import_data(data: { override_params: { description: "Overridden" } }) - expect(restored_project_json).to eq(true) - expect(project.lfs_enabled).to be_falsey - end + expect(restored_project_json).to eq(true) + expect(project.description).to eq("Overridden") + end - it 'overrides project feature access levels' do - access_level_keys = project.project_feature.attributes.keys.select { |a| a =~ /_access_level/ } + it 'does not allow setting params that are excluded from import_export settings' do + project.create_import_data(data: { override_params: { lfs_enabled: true } }) - # `pages_access_level` is not included, since it is not available in the public API - # and has a dependency on project's visibility level - # see ProjectFeature model - access_level_keys.delete('pages_access_level') + expect(restored_project_json).to eq(true) + expect(project.lfs_enabled).to be_falsey + end + + it 'overrides project feature access levels' do + access_level_keys = project.project_feature.attributes.keys.select { |a| a =~ /_access_level/ } + + # `pages_access_level` is not included, since it is not available in the public API + # and has a dependency on project's visibility level + # see ProjectFeature model + access_level_keys.delete('pages_access_level') - disabled_access_levels = Hash[access_level_keys.collect { |item| [item, 'disabled'] }] + disabled_access_levels = Hash[access_level_keys.collect { |item| [item, 'disabled'] }] - project.create_import_data(data: { override_params: disabled_access_levels }) + project.create_import_data(data: { override_params: disabled_access_levels }) - expect(restored_project_json).to eq(true) + expect(restored_project_json).to eq(true) - aggregate_failures do - access_level_keys.each do |key| - expect(project.public_send(key)).to eq(ProjectFeature::DISABLED) + aggregate_failures do + access_level_keys.each do |key| + expect(project.public_send(key)).to eq(ProjectFeature::DISABLED) + end end end end - end - context 'with a project that has a group' do - let!(:project) do - create(:project, - :builds_disabled, - :issues_disabled, - name: 'project', - path: 'project', - group: create(:group, visibility_level: Gitlab::VisibilityLevel::PRIVATE)) - end + context 'with a project that has a group' do + let!(:project) do + create(:project, + :builds_disabled, + :issues_disabled, + name: 'project', + path: 'project', + group: create(:group, visibility_level: Gitlab::VisibilityLevel::PRIVATE)) + end - before do - setup_import_export_config('group') - expect(restored_project_json).to eq(true) - end + before do + setup_import_export_config('group') + setup_reader(reader) - it_behaves_like 'restores project successfully', - issues: 3, - labels: 2, - label_with_priorities: 'A project label', - milestones: 2, - first_issue_labels: 1 - - it_behaves_like 'restores group correctly', - labels: 0, - milestones: 0, - first_issue_labels: 1 - - it 'restores issue states' do - expect(project.issues.with_state(:closed).count).to eq(1) - expect(project.issues.with_state(:opened).count).to eq(2) - end - end + expect(restored_project_json).to eq(true) + end - context 'with existing group models' do - let!(:project) do - create(:project, - :builds_disabled, - :issues_disabled, - name: 'project', - path: 'project', - group: create(:group)) - end + after do + cleanup_artifacts_from_extract_archive('group') + end - before do - setup_import_export_config('light') - end + it_behaves_like 'restores project successfully', + issues: 3, + labels: 2, + label_with_priorities: 'A project label', + milestones: 2, + first_issue_labels: 1 - it 'does not import any templated services' do - expect(restored_project_json).to eq(true) + it_behaves_like 'restores group correctly', + labels: 0, + milestones: 0, + first_issue_labels: 1 - expect(project.services.where(template: true).count).to eq(0) + it 'restores issue states' do + expect(project.issues.with_state(:closed).count).to eq(1) + expect(project.issues.with_state(:opened).count).to eq(2) + end end - it 'does not import any instance services' do - expect(restored_project_json).to eq(true) + context 'with existing group models' do + let!(:project) do + create(:project, + :builds_disabled, + :issues_disabled, + name: 'project', + path: 'project', + group: create(:group)) + end - expect(project.services.where(instance: true).count).to eq(0) - end + before do + setup_import_export_config('light') + setup_reader(reader) + end - it 'imports labels' do - create(:group_label, name: 'Another label', group: project.group) + after do + cleanup_artifacts_from_extract_archive('light') + end - expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error) + it 'does not import any templated services' do + expect(restored_project_json).to eq(true) - expect(restored_project_json).to eq(true) - expect(project.labels.count).to eq(1) - end + expect(project.services.where(template: true).count).to eq(0) + end - it 'imports milestones' do - create(:milestone, name: 'A milestone', group: project.group) + it 'does not import any instance services' do + expect(restored_project_json).to eq(true) - expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error) + expect(project.services.where(instance: true).count).to eq(0) + end - expect(restored_project_json).to eq(true) - expect(project.group.milestones.count).to eq(1) - expect(project.milestones.count).to eq(0) - end - end + it 'imports labels' do + create(:group_label, name: 'Another label', group: project.group) - context 'with clashing milestones on IID' do - let!(:project) do - create(:project, - :builds_disabled, - :issues_disabled, - name: 'project', - path: 'project', - group: create(:group)) - end + expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error) - before do - setup_import_export_config('milestone-iid') - end + expect(restored_project_json).to eq(true) + expect(project.labels.count).to eq(1) + end - it 'preserves the project milestone IID' do - expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error) + it 'imports milestones' do + create(:milestone, name: 'A milestone', group: project.group) - expect(restored_project_json).to eq(true) - expect(project.milestones.count).to eq(2) - expect(Milestone.find_by_title('Another milestone').iid).to eq(1) - expect(Milestone.find_by_title('Group-level milestone').iid).to eq(2) - end - end + expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error) - context 'with external authorization classification labels' do - before do - setup_import_export_config('light') + expect(restored_project_json).to eq(true) + expect(project.group.milestones.count).to eq(1) + expect(project.milestones.count).to eq(0) + end end - it 'converts empty external classification authorization labels to nil' do - project.create_import_data(data: { override_params: { external_authorization_classification_label: "" } }) + context 'with clashing milestones on IID' do + let!(:project) do + create(:project, + :builds_disabled, + :issues_disabled, + name: 'project', + path: 'project', + group: create(:group)) + end - expect(restored_project_json).to eq(true) - expect(project.external_authorization_classification_label).to be_nil - end + before do + setup_import_export_config('milestone-iid') + setup_reader(reader) + end - it 'preserves valid external classification authorization labels' do - project.create_import_data(data: { override_params: { external_authorization_classification_label: "foobar" } }) + after do + cleanup_artifacts_from_extract_archive('milestone-iid') + end - expect(restored_project_json).to eq(true) - expect(project.external_authorization_classification_label).to eq("foobar") - end - end - end + it 'preserves the project milestone IID' do + expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error) - context 'Minimal JSON' do - let(:project) { create(:project) } - let(:user) { create(:user) } - let(:tree_hash) { { 'visibility_level' => visibility } } - let(:restorer) do - described_class.new(user: user, shared: shared, project: project) - end + expect(restored_project_json).to eq(true) + expect(project.milestones.count).to eq(2) + expect(Milestone.find_by_title('Another milestone').iid).to eq(1) + expect(Milestone.find_by_title('Group-level milestone').iid).to eq(2) + end + end - before do - allow_any_instance_of(Gitlab::ImportExport::JSON::LegacyReader::File).to receive(:valid?).and_return(true) - allow_any_instance_of(Gitlab::ImportExport::JSON::LegacyReader::File).to receive(:tree_hash) { tree_hash } - end + context 'with external authorization classification labels' do + before do + setup_import_export_config('light') + setup_reader(reader) + end - context 'no group visibility' do - let(:visibility) { Gitlab::VisibilityLevel::PRIVATE } + after do + cleanup_artifacts_from_extract_archive('light') + end - it 'uses the project visibility' do - expect(restorer.restore).to eq(true) - expect(restorer.project.visibility_level).to eq(visibility) - end - end + it 'converts empty external classification authorization labels to nil' do + project.create_import_data(data: { override_params: { external_authorization_classification_label: "" } }) - context 'with restricted internal visibility' do - describe 'internal project' do - let(:visibility) { Gitlab::VisibilityLevel::INTERNAL } + expect(restored_project_json).to eq(true) + expect(project.external_authorization_classification_label).to be_nil + end - it 'uses private visibility' do - stub_application_setting(restricted_visibility_levels: [Gitlab::VisibilityLevel::INTERNAL]) + it 'preserves valid external classification authorization labels' do + project.create_import_data(data: { override_params: { external_authorization_classification_label: "foobar" } }) - expect(restorer.restore).to eq(true) - expect(restorer.project.visibility_level).to eq(Gitlab::VisibilityLevel::PRIVATE) + expect(restored_project_json).to eq(true) + expect(project.external_authorization_classification_label).to eq("foobar") end end end - context 'with group visibility' do - before do - group = create(:group, visibility_level: group_visibility) - - project.update(group: group) + context 'Minimal JSON' do + let(:project) { create(:project) } + let(:user) { create(:user) } + let(:tree_hash) { { 'visibility_level' => visibility } } + let(:restorer) do + described_class.new(user: user, shared: shared, project: project) end - context 'private group visibility' do - let(:group_visibility) { Gitlab::VisibilityLevel::PRIVATE } - let(:visibility) { Gitlab::VisibilityLevel::PUBLIC } - - it 'uses the group visibility' do - expect(restorer.restore).to eq(true) - expect(restorer.project.visibility_level).to eq(group_visibility) - end + before do + allow_any_instance_of(Gitlab::ImportExport::JSON::LegacyReader::File).to receive(:exist?).and_return(true) + allow_any_instance_of(Gitlab::ImportExport::JSON::NdjsonReader).to receive(:exist?).and_return(false) + allow_any_instance_of(Gitlab::ImportExport::JSON::LegacyReader::File).to receive(:tree_hash) { tree_hash } end - context 'public group visibility' do - let(:group_visibility) { Gitlab::VisibilityLevel::PUBLIC } + context 'no group visibility' do let(:visibility) { Gitlab::VisibilityLevel::PRIVATE } it 'uses the project visibility' do @@ -871,17 +894,11 @@ describe Gitlab::ImportExport::Project::TreeRestorer do end end - context 'internal group visibility' do - let(:group_visibility) { Gitlab::VisibilityLevel::INTERNAL } - let(:visibility) { Gitlab::VisibilityLevel::PUBLIC } - - it 'uses the group visibility' do - expect(restorer.restore).to eq(true) - expect(restorer.project.visibility_level).to eq(group_visibility) - end + context 'with restricted internal visibility' do + describe 'internal project' do + let(:visibility) { Gitlab::VisibilityLevel::INTERNAL } - context 'with restricted internal visibility' do - it 'sets private visibility' do + it 'uses private visibility' do stub_application_setting(restricted_visibility_levels: [Gitlab::VisibilityLevel::INTERNAL]) expect(restorer.restore).to eq(true) @@ -889,43 +906,116 @@ describe Gitlab::ImportExport::Project::TreeRestorer do end end end - end - end - context 'JSON with invalid records' do - subject(:restored_project_json) { project_tree_restorer.restore } + context 'with group visibility' do + before do + group = create(:group, visibility_level: group_visibility) + + project.update(group: group) + end - let(:user) { create(:user) } - let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') } - let(:project_tree_restorer) { described_class.new(user: user, shared: shared, project: project) } + context 'private group visibility' do + let(:group_visibility) { Gitlab::VisibilityLevel::PRIVATE } + let(:visibility) { Gitlab::VisibilityLevel::PUBLIC } - before do - setup_import_export_config('with_invalid_records') + it 'uses the group visibility' do + expect(restorer.restore).to eq(true) + expect(restorer.project.visibility_level).to eq(group_visibility) + end + end + + context 'public group visibility' do + let(:group_visibility) { Gitlab::VisibilityLevel::PUBLIC } + let(:visibility) { Gitlab::VisibilityLevel::PRIVATE } + + it 'uses the project visibility' do + expect(restorer.restore).to eq(true) + expect(restorer.project.visibility_level).to eq(visibility) + end + end - subject + context 'internal group visibility' do + let(:group_visibility) { Gitlab::VisibilityLevel::INTERNAL } + let(:visibility) { Gitlab::VisibilityLevel::PUBLIC } + + it 'uses the group visibility' do + expect(restorer.restore).to eq(true) + expect(restorer.project.visibility_level).to eq(group_visibility) + end + + context 'with restricted internal visibility' do + it 'sets private visibility' do + stub_application_setting(restricted_visibility_levels: [Gitlab::VisibilityLevel::INTERNAL]) + + expect(restorer.restore).to eq(true) + expect(restorer.project.visibility_level).to eq(Gitlab::VisibilityLevel::PRIVATE) + end + end + end + end end - context 'when failures occur because a relation fails to be processed' do - it_behaves_like 'restores project successfully', - issues: 0, - labels: 0, - label_with_priorities: nil, - milestones: 1, - first_issue_labels: 0, - services: 0, - import_failures: 1 - - it 'records the failures in the database' do - import_failure = ImportFailure.last - - expect(import_failure.project_id).to eq(project.id) - expect(import_failure.relation_key).to eq('milestones') - expect(import_failure.relation_index).to be_present - expect(import_failure.exception_class).to eq('ActiveRecord::RecordInvalid') - expect(import_failure.exception_message).to be_present - expect(import_failure.correlation_id_value).not_to be_empty - expect(import_failure.created_at).to be_present + context 'JSON with invalid records' do + subject(:restored_project_json) { project_tree_restorer.restore } + + let(:user) { create(:user) } + let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') } + let(:project_tree_restorer) { described_class.new(user: user, shared: shared, project: project) } + + before do + setup_import_export_config('with_invalid_records') + setup_reader(reader) + + subject + end + + after do + cleanup_artifacts_from_extract_archive('with_invalid_records') end + + context 'when failures occur because a relation fails to be processed' do + it_behaves_like 'restores project successfully', + issues: 0, + labels: 0, + label_with_priorities: nil, + milestones: 1, + first_issue_labels: 0, + services: 0, + import_failures: 1 + + it 'records the failures in the database' do + import_failure = ImportFailure.last + + expect(import_failure.project_id).to eq(project.id) + expect(import_failure.relation_key).to eq('milestones') + expect(import_failure.relation_index).to be_present + expect(import_failure.exception_class).to eq('ActiveRecord::RecordInvalid') + expect(import_failure.exception_message).to be_present + expect(import_failure.correlation_id_value).not_to be_empty + expect(import_failure.created_at).to be_present + end + end + end + end + + context 'enable ndjson import' do + before_all do + # Test suite `restore project tree` run `project_tree_restorer.restore` in `before_all`. + # `Enable all features by default for testing` happens in `before(:each)` + # So it requires manually enable feature flag to allow ndjson_reader + Feature.enable(:project_import_ndjson) + end + + it_behaves_like 'project tree restorer work properly', :legacy_reader + + it_behaves_like 'project tree restorer work properly', :ndjson_reader + end + + context 'disable ndjson import' do + before do + stub_feature_flags(project_import_ndjson: false) end + + it_behaves_like 'project tree restorer work properly', :legacy_reader end end diff --git a/spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb b/spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb index 52e1efa70e0..0b58a75220d 100644 --- a/spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb +++ b/spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb @@ -14,7 +14,7 @@ describe Gitlab::ImportExport::RelationTreeRestorer do let(:user) { create(:user) } let(:shared) { Gitlab::ImportExport::Shared.new(importable) } - let(:attributes) { {} } + let(:attributes) { relation_reader.consume_attributes(importable_name) } let(:members_mapper) do Gitlab::ImportExport::MembersMapper.new(exported_members: {}, user: user, importable: importable) @@ -30,7 +30,7 @@ describe Gitlab::ImportExport::RelationTreeRestorer do relation_factory: relation_factory, reader: reader, importable: importable, - importable_path: nil, + importable_path: importable_path, importable_attributes: attributes ) end @@ -94,21 +94,24 @@ describe Gitlab::ImportExport::RelationTreeRestorer do end context 'when restoring a project' do - let(:path) { 'spec/fixtures/lib/gitlab/import_export/complex/project.json' } let(:importable) { create(:project, :builds_enabled, :issues_disabled, name: 'project', path: 'project') } + let(:importable_name) { 'project' } + let(:importable_path) { 'project' } let(:object_builder) { Gitlab::ImportExport::Project::ObjectBuilder } let(:relation_factory) { Gitlab::ImportExport::Project::RelationFactory } let(:reader) { Gitlab::ImportExport::Reader.new(shared: shared) } context 'using legacy reader' do + let(:path) { 'spec/fixtures/lib/gitlab/import_export/complex/project.json' } let(:relation_reader) do Gitlab::ImportExport::JSON::LegacyReader::File.new( path, - relation_names: reader.project_relation_names + relation_names: reader.project_relation_names, + allowed_path: 'project' ) end - let(:attributes) { relation_reader.consume_attributes(nil) } + let(:attributes) { relation_reader.consume_attributes('project') } it_behaves_like 'import project successfully' @@ -118,6 +121,21 @@ describe Gitlab::ImportExport::RelationTreeRestorer do include_examples 'logging of relations creation' end + + context 'using ndjson reader' do + let(:path) { 'spec/fixtures/lib/gitlab/import_export/complex/tree' } + let(:relation_reader) { Gitlab::ImportExport::JSON::NdjsonReader.new(path) } + + before :all do + extract_archive('spec/fixtures/lib/gitlab/import_export/complex', 'tree.tar.gz') + end + + after :all do + cleanup_artifacts_from_extract_archive('complex') + end + + it_behaves_like 'import project successfully' + end end end @@ -125,9 +143,16 @@ describe Gitlab::ImportExport::RelationTreeRestorer do let(:path) { 'spec/fixtures/lib/gitlab/import_export/group_exports/no_children/group.json' } let(:group) { create(:group) } let(:importable) { create(:group, parent: group) } + let(:importable_name) { nil } + let(:importable_path) { nil } let(:object_builder) { Gitlab::ImportExport::Group::ObjectBuilder } let(:relation_factory) { Gitlab::ImportExport::Group::RelationFactory } - let(:relation_reader) { Gitlab::ImportExport::JSON::LegacyReader::File.new(path, relation_names: reader.group_relation_names) } + let(:relation_reader) do + Gitlab::ImportExport::JSON::LegacyReader::File.new( + path, + relation_names: reader.group_relation_names) + end + let(:reader) do Gitlab::ImportExport::Reader.new( shared: shared, @@ -135,6 +160,10 @@ describe Gitlab::ImportExport::RelationTreeRestorer do ) end + it 'restores group tree' do + expect(subject).to eq(true) + end + include_examples 'logging of relations creation' end end |