summaryrefslogtreecommitdiff
path: root/spec/lib
diff options
context:
space:
mode:
authorGitLab Bot <gitlab-bot@gitlab.com>2020-04-08 03:09:31 +0000
committerGitLab Bot <gitlab-bot@gitlab.com>2020-04-08 03:09:31 +0000
commite2ee1eec50aa8df8543d7ecc585ec0ba5ee544ac (patch)
tree7998650d27ada12ee7d06a21cbb3b5e89f298378 /spec/lib
parent060c842402c00f830a810702600cbe39dfa6cf62 (diff)
downloadgitlab-ce-e2ee1eec50aa8df8543d7ecc585ec0ba5ee544ac.tar.gz
Add latest changes from gitlab-org/gitlab@master
Diffstat (limited to 'spec/lib')
-rw-r--r--spec/lib/gitlab/graphql/pagination/externally_paginated_array_connection_spec.rb (renamed from spec/lib/gitlab/graphql/connections/externally_paginated_array_connection_spec.rb)18
-rw-r--r--spec/lib/gitlab/graphql/pagination/filterable_array_connection_spec.rb (renamed from spec/lib/gitlab/graphql/connections/filterable_array_connection_spec.rb)12
-rw-r--r--spec/lib/gitlab/graphql/pagination/keyset/conditions/not_null_condition_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/conditions/not_null_condition_spec.rb)4
-rw-r--r--spec/lib/gitlab/graphql/pagination/keyset/conditions/null_condition_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/conditions/null_condition_spec.rb)4
-rw-r--r--spec/lib/gitlab/graphql/pagination/keyset/connection_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/connection_spec.rb)19
-rw-r--r--spec/lib/gitlab/graphql/pagination/keyset/order_info_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/order_info_spec.rb)2
-rw-r--r--spec/lib/gitlab/graphql/pagination/keyset/query_builder_spec.rb (renamed from spec/lib/gitlab/graphql/connections/keyset/query_builder_spec.rb)4
-rw-r--r--spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb70
-rw-r--r--spec/lib/gitlab/import_export/json/ndjson_reader_spec.rb113
-rw-r--r--spec/lib/gitlab/import_export/project/tree_restorer_spec.rb1390
-rw-r--r--spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb41
11 files changed, 948 insertions, 729 deletions
diff --git a/spec/lib/gitlab/graphql/connections/externally_paginated_array_connection_spec.rb b/spec/lib/gitlab/graphql/pagination/externally_paginated_array_connection_spec.rb
index 83c94ed6260..85a5b1dacc7 100644
--- a/spec/lib/gitlab/graphql/connections/externally_paginated_array_connection_spec.rb
+++ b/spec/lib/gitlab/graphql/pagination/externally_paginated_array_connection_spec.rb
@@ -2,7 +2,7 @@
require 'spec_helper'
-describe Gitlab::Graphql::Connections::ExternallyPaginatedArrayConnection do
+describe Gitlab::Graphql::Pagination::ExternallyPaginatedArrayConnection do
let(:prev_cursor) { 1 }
let(:next_cursor) { 6 }
let(:values) { [2, 3, 4, 5] }
@@ -10,21 +10,13 @@ describe Gitlab::Graphql::Connections::ExternallyPaginatedArrayConnection do
let(:arguments) { {} }
subject(:connection) do
- described_class.new(all_nodes, arguments)
+ described_class.new(all_nodes, { max_page_size: values.size }.merge(arguments))
end
- describe '#sliced_nodes' do
- let(:sliced_nodes) { connection.sliced_nodes }
+ describe '#nodes' do
+ let(:paged_nodes) { connection.nodes }
- it 'returns all the nodes' do
- expect(connection.sliced_nodes).to eq(values)
- end
- end
-
- describe '#paged_nodes' do
- let(:paged_nodes) { connection.send(:paged_nodes) }
-
- it_behaves_like "connection with paged nodes" do
+ it_behaves_like 'connection with paged nodes' do
let(:paged_nodes_size) { values.size }
end
end
diff --git a/spec/lib/gitlab/graphql/connections/filterable_array_connection_spec.rb b/spec/lib/gitlab/graphql/pagination/filterable_array_connection_spec.rb
index b2f0862be62..c82e3ad3019 100644
--- a/spec/lib/gitlab/graphql/connections/filterable_array_connection_spec.rb
+++ b/spec/lib/gitlab/graphql/pagination/filterable_array_connection_spec.rb
@@ -2,19 +2,19 @@
require 'spec_helper'
-describe Gitlab::Graphql::Connections::FilterableArrayConnection do
+describe Gitlab::Graphql::Pagination::FilterableArrayConnection do
let(:callback) { proc { |nodes| nodes } }
let(:all_nodes) { Gitlab::Graphql::FilterableArray.new(callback, 1, 2, 3, 4, 5) }
let(:arguments) { {} }
subject(:connection) do
- described_class.new(all_nodes, arguments, max_page_size: 3)
+ described_class.new(all_nodes, { max_page_size: 3 }.merge(arguments))
end
- describe '#paged_nodes' do
- let(:paged_nodes) { subject.paged_nodes }
+ describe '#nodes' do
+ let(:paged_nodes) { subject.nodes }
- it_behaves_like "connection with paged nodes" do
+ it_behaves_like 'connection with paged nodes' do
let(:paged_nodes_size) { 3 }
end
@@ -22,7 +22,7 @@ describe Gitlab::Graphql::Connections::FilterableArrayConnection do
let(:callback) { proc { |nodes| nodes[1..-1] } }
it 'does not return filtered elements' do
- expect(subject.paged_nodes).to contain_exactly(all_nodes[1], all_nodes[2])
+ expect(subject.nodes).to contain_exactly(all_nodes[1], all_nodes[2])
end
end
end
diff --git a/spec/lib/gitlab/graphql/connections/keyset/conditions/not_null_condition_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/conditions/not_null_condition_spec.rb
index 26fc5344871..33dc6bdea97 100644
--- a/spec/lib/gitlab/graphql/connections/keyset/conditions/not_null_condition_spec.rb
+++ b/spec/lib/gitlab/graphql/pagination/keyset/conditions/not_null_condition_spec.rb
@@ -2,7 +2,7 @@
require 'spec_helper'
-describe Gitlab::Graphql::Connections::Keyset::Conditions::NotNullCondition do
+describe Gitlab::Graphql::Pagination::Keyset::Conditions::NotNullCondition do
describe '#build' do
let(:operators) { ['>', '>'] }
let(:before_or_after) { :after }
@@ -75,7 +75,7 @@ describe Gitlab::Graphql::Connections::Keyset::Conditions::NotNullCondition do
context 'when ordering by LOWER' do
let(:arel_table) { Project.arel_table }
let(:relation) { Project.order(arel_table['name'].lower.asc).order(:id) }
- let(:order_list) { Gitlab::Graphql::Connections::Keyset::OrderInfo.build_order_list(relation) }
+ let(:order_list) { Gitlab::Graphql::Pagination::Keyset::OrderInfo.build_order_list(relation) }
let(:values) { ['Test', 500] }
context 'when :after' do
diff --git a/spec/lib/gitlab/graphql/connections/keyset/conditions/null_condition_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/conditions/null_condition_spec.rb
index be0a21b2438..1c74f2fb0ab 100644
--- a/spec/lib/gitlab/graphql/connections/keyset/conditions/null_condition_spec.rb
+++ b/spec/lib/gitlab/graphql/pagination/keyset/conditions/null_condition_spec.rb
@@ -2,7 +2,7 @@
require 'spec_helper'
-describe Gitlab::Graphql::Connections::Keyset::Conditions::NullCondition do
+describe Gitlab::Graphql::Pagination::Keyset::Conditions::NullCondition do
describe '#build' do
let(:values) { [nil, 500] }
let(:operators) { [nil, '>'] }
@@ -58,7 +58,7 @@ describe Gitlab::Graphql::Connections::Keyset::Conditions::NullCondition do
context 'when ordering by LOWER' do
let(:arel_table) { Project.arel_table }
let(:relation) { Project.order(arel_table['name'].lower.asc).order(:id) }
- let(:order_list) { Gitlab::Graphql::Connections::Keyset::OrderInfo.build_order_list(relation) }
+ let(:order_list) { Gitlab::Graphql::Pagination::Keyset::OrderInfo.build_order_list(relation) }
context 'when :after' do
it 'generates sql' do
diff --git a/spec/lib/gitlab/graphql/connections/keyset/connection_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/connection_spec.rb
index c193ab2b50f..fdacecbaca6 100644
--- a/spec/lib/gitlab/graphql/connections/keyset/connection_spec.rb
+++ b/spec/lib/gitlab/graphql/pagination/keyset/connection_spec.rb
@@ -2,25 +2,28 @@
require 'spec_helper'
-describe Gitlab::Graphql::Connections::Keyset::Connection do
+describe Gitlab::Graphql::Pagination::Keyset::Connection do
let(:nodes) { Project.all.order(id: :asc) }
let(:arguments) { {} }
+ let(:query_type) { GraphQL::ObjectType.new }
+ let(:schema) { GraphQL::Schema.define(query: query_type, mutation: nil)}
+ let(:context) { GraphQL::Query::Context.new(query: OpenStruct.new(schema: schema), values: nil, object: nil) }
subject(:connection) do
- described_class.new(nodes, arguments, max_page_size: 3)
+ described_class.new(nodes, { context: context, max_page_size: 3 }.merge(arguments))
end
def encoded_cursor(node)
- described_class.new(nodes, {}).cursor_from_node(node)
+ described_class.new(nodes, { context: context }).cursor_for(node)
end
def decoded_cursor(cursor)
JSON.parse(Base64Bp.urlsafe_decode64(cursor))
end
- describe '#cursor_from_nodes' do
+ describe '#cursor_for' do
let(:project) { create(:project) }
- let(:cursor) { connection.cursor_from_node(project) }
+ let(:cursor) { connection.cursor_for(project) }
it 'returns an encoded ID' do
expect(decoded_cursor(cursor)).to eq('id' => project.id.to_s)
@@ -264,11 +267,11 @@ describe Gitlab::Graphql::Connections::Keyset::Connection do
end
end
- describe '#paged_nodes' do
+ describe '#nodes' do
let_it_be(:all_nodes) { create_list(:project, 5) }
- let(:paged_nodes) { subject.paged_nodes }
+ let(:paged_nodes) { subject.nodes }
- it_behaves_like "connection with paged nodes" do
+ it_behaves_like 'connection with paged nodes' do
let(:paged_nodes_size) { 3 }
end
diff --git a/spec/lib/gitlab/graphql/connections/keyset/order_info_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/order_info_spec.rb
index eb823fc0122..cb5656d7c00 100644
--- a/spec/lib/gitlab/graphql/connections/keyset/order_info_spec.rb
+++ b/spec/lib/gitlab/graphql/pagination/keyset/order_info_spec.rb
@@ -2,7 +2,7 @@
require 'spec_helper'
-describe Gitlab::Graphql::Connections::Keyset::OrderInfo do
+describe Gitlab::Graphql::Pagination::Keyset::OrderInfo do
describe '#build_order_list' do
let(:order_list) { described_class.build_order_list(relation) }
diff --git a/spec/lib/gitlab/graphql/connections/keyset/query_builder_spec.rb b/spec/lib/gitlab/graphql/pagination/keyset/query_builder_spec.rb
index b46ce4bf023..7fd5b889cf9 100644
--- a/spec/lib/gitlab/graphql/connections/keyset/query_builder_spec.rb
+++ b/spec/lib/gitlab/graphql/pagination/keyset/query_builder_spec.rb
@@ -2,7 +2,7 @@
require 'spec_helper'
-describe Gitlab::Graphql::Connections::Keyset::QueryBuilder do
+describe Gitlab::Graphql::Pagination::Keyset::QueryBuilder do
context 'when number of ordering fields is 0' do
it 'raises an error' do
expect { described_class.new(Issue.arel_table, [], {}, :after) }
@@ -12,7 +12,7 @@ describe Gitlab::Graphql::Connections::Keyset::QueryBuilder do
describe '#conditions' do
let(:relation) { Issue.order(relative_position: :desc).order(:id) }
- let(:order_list) { Gitlab::Graphql::Connections::Keyset::OrderInfo.build_order_list(relation) }
+ let(:order_list) { Gitlab::Graphql::Pagination::Keyset::OrderInfo.build_order_list(relation) }
let(:arel_table) { Issue.arel_table }
let(:builder) { described_class.new(arel_table, order_list, decoded_cursor, before_or_after) }
let(:before_or_after) { :after }
diff --git a/spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb b/spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb
index 297a5946703..3e9bd3fe741 100644
--- a/spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb
+++ b/spec/lib/gitlab/import_export/json/legacy_reader/shared_example.rb
@@ -15,7 +15,6 @@ RSpec.shared_examples 'import/export json legacy reader' do
subject { legacy_reader.consume_attributes("project") }
context 'no excluded attributes' do
- let(:excluded_attributes) { [] }
let(:relation_names) { [] }
it 'returns the whole tree from parsed JSON' do
@@ -42,60 +41,53 @@ RSpec.shared_examples 'import/export json legacy reader' do
describe '#consume_relation' do
context 'when valid path is passed' do
- let(:key) { 'description' }
+ let(:key) { 'labels' }
- context 'block not given' do
- it 'returns value of the key' do
- expect(legacy_reader).to receive(:relations).and_return({ key => 'test value' })
- expect(legacy_reader.consume_relation("project", key)).to eq('test value')
- end
- end
+ subject { legacy_reader.consume_relation("project", key) }
- context 'key has been consumed' do
- before do
- legacy_reader.consume_relation("project", key)
+ context 'key has not been consumed' do
+ it 'returns an Enumerator' do
+ expect(subject).to be_an_instance_of(Enumerator)
end
- it 'does not yield' do
- expect do |blk|
- legacy_reader.consume_relation("project", key, &blk)
- end.not_to yield_control
- end
- end
+ context 'value is nil' do
+ before do
+ expect(legacy_reader).to receive(:relations).and_return({ key => nil })
+ end
- context 'value is nil' do
- before do
- expect(legacy_reader).to receive(:relations).and_return({ key => nil })
+ it 'yields nothing to the Enumerator' do
+ expect(subject.to_a).to eq([])
+ end
end
- it 'does not yield' do
- expect do |blk|
- legacy_reader.consume_relation("project", key, &blk)
- end.not_to yield_control
- end
- end
+ context 'value is an array' do
+ before do
+ expect(legacy_reader).to receive(:relations).and_return({ key => %w[label1 label2] })
+ end
- context 'value is not array' do
- before do
- expect(legacy_reader).to receive(:relations).and_return({ key => 'value' })
+ it 'yields every relation value to the Enumerator' do
+ expect(subject.to_a).to eq([['label1', 0], ['label2', 1]])
+ end
end
- it 'yield the value with index 0' do
- expect do |blk|
- legacy_reader.consume_relation("project", key, &blk)
- end.to yield_with_args('value', 0)
+ context 'value is not array' do
+ before do
+ expect(legacy_reader).to receive(:relations).and_return({ key => 'non-array value' })
+ end
+
+ it 'yields the value with index 0 to the Enumerator' do
+ expect(subject.to_a).to eq([['non-array value', 0]])
+ end
end
end
- context 'value is an array' do
+ context 'key has been consumed' do
before do
- expect(legacy_reader).to receive(:relations).and_return({ key => %w[item1 item2 item3] })
+ legacy_reader.consume_relation("project", key).first
end
- it 'yield each array element with index' do
- expect do |blk|
- legacy_reader.consume_relation("project", key, &blk)
- end.to yield_successive_args(['item1', 0], ['item2', 1], ['item3', 2])
+ it 'yields nothing to the Enumerator' do
+ expect(subject.to_a).to eq([])
end
end
end
diff --git a/spec/lib/gitlab/import_export/json/ndjson_reader_spec.rb b/spec/lib/gitlab/import_export/json/ndjson_reader_spec.rb
new file mode 100644
index 00000000000..40b784fdb87
--- /dev/null
+++ b/spec/lib/gitlab/import_export/json/ndjson_reader_spec.rb
@@ -0,0 +1,113 @@
+# frozen_string_literal: true
+
+require 'spec_helper'
+
+describe Gitlab::ImportExport::JSON::NdjsonReader do
+ include ImportExport::CommonUtil
+
+ let(:fixture) { 'spec/fixtures/lib/gitlab/import_export/light/tree' }
+ let(:root_tree) { JSON.parse(File.read(File.join(fixture, 'project.json'))) }
+ let(:ndjson_reader) { described_class.new(dir_path) }
+ let(:importable_path) { 'project' }
+
+ before :all do
+ extract_archive('spec/fixtures/lib/gitlab/import_export/light', 'tree.tar.gz')
+ end
+
+ after :all do
+ cleanup_artifacts_from_extract_archive('light')
+ end
+
+ describe '#exist?' do
+ subject { ndjson_reader.exist? }
+
+ context 'given valid dir_path' do
+ let(:dir_path) { fixture }
+
+ it { is_expected.to be true }
+ end
+
+ context 'given invalid dir_path' do
+ let(:dir_path) { 'invalid-dir-path' }
+
+ it { is_expected.to be false }
+ end
+ end
+
+ describe '#legacy?' do
+ let(:dir_path) { fixture }
+
+ subject { ndjson_reader.legacy? }
+
+ it { is_expected.to be false }
+ end
+
+ describe '#consume_attributes' do
+ let(:dir_path) { fixture }
+
+ subject { ndjson_reader.consume_attributes(importable_path) }
+
+ it 'returns the whole root tree from parsed JSON' do
+ expect(subject).to eq(root_tree)
+ end
+ end
+
+ describe '#consume_relation' do
+ let(:dir_path) { fixture }
+
+ subject { ndjson_reader.consume_relation(importable_path, key) }
+
+ context 'given any key' do
+ let(:key) { 'any-key' }
+
+ it 'returns an Enumerator' do
+ expect(subject).to be_an_instance_of(Enumerator)
+ end
+ end
+
+ context 'key has been consumed' do
+ let(:key) { 'issues' }
+
+ before do
+ ndjson_reader.consume_relation(importable_path, key).first
+ end
+
+ it 'yields nothing to the Enumerator' do
+ expect(subject.to_a).to eq([])
+ end
+ end
+
+ context 'key has not been consumed' do
+ context 'relation file does not exist' do
+ let(:key) { 'non-exist-relation-file-name' }
+
+ before do
+ relation_file_path = File.join(dir_path, importable_path, "#{key}.ndjson")
+ expect(File).to receive(:exist?).with(relation_file_path).and_return(false)
+ end
+
+ it 'yields nothing to the Enumerator' do
+ expect(subject.to_a).to eq([])
+ end
+ end
+
+ context 'relation file is empty' do
+ let(:key) { 'empty' }
+
+ it 'yields nothing to the Enumerator' do
+ expect(subject.to_a).to eq([])
+ end
+ end
+
+ context 'relation file contains multiple lines' do
+ let(:key) { 'custom_attributes' }
+ let(:attr_1) { JSON.parse('{"id":201,"project_id":5,"created_at":"2016-06-14T15:01:51.315Z","updated_at":"2016-06-14T15:01:51.315Z","key":"color","value":"red"}') }
+ let(:attr_2) { JSON.parse('{"id":202,"project_id":5,"created_at":"2016-06-14T15:01:51.315Z","updated_at":"2016-06-14T15:01:51.315Z","key":"size","value":"small"}') }
+
+ it 'yields every relation value to the Enumerator' do
+ expect(subject.to_a).to eq([[attr_1, 0], [attr_2, 1]])
+ end
+ end
+ end
+ end
+end
diff --git a/spec/lib/gitlab/import_export/project/tree_restorer_spec.rb b/spec/lib/gitlab/import_export/project/tree_restorer_spec.rb
index 247d455fecc..96aed774cfc 100644
--- a/spec/lib/gitlab/import_export/project/tree_restorer_spec.rb
+++ b/spec/lib/gitlab/import_export/project/tree_restorer_spec.rb
@@ -11,76 +11,83 @@ describe Gitlab::ImportExport::Project::TreeRestorer do
let(:shared) { project.import_export_shared }
- describe 'restore project tree' do
- before_all do
- # Using an admin for import, so we can check assignment of existing members
- @user = create(:admin)
- @existing_members = [
- create(:user, email: 'bernard_willms@gitlabexample.com'),
- create(:user, email: 'saul_will@gitlabexample.com')
- ]
+ RSpec.shared_examples 'project tree restorer work properly' do |reader|
+ describe 'restore project tree' do
+ before_all do
+ # Using an admin for import, so we can check assignment of existing members
+ @user = create(:admin)
+ @existing_members = [
+ create(:user, email: 'bernard_willms@gitlabexample.com'),
+ create(:user, email: 'saul_will@gitlabexample.com')
+ ]
- RSpec::Mocks.with_temporary_scope do
- @project = create(:project, :builds_enabled, :issues_disabled, name: 'project', path: 'project')
- @shared = @project.import_export_shared
+ RSpec::Mocks.with_temporary_scope do
+ @project = create(:project, :builds_enabled, :issues_disabled, name: 'project', path: 'project')
+ @shared = @project.import_export_shared
- setup_import_export_config('complex')
+ setup_import_export_config('complex')
+ setup_reader(reader)
- allow_any_instance_of(Repository).to receive(:fetch_source_branch!).and_return(true)
- allow_any_instance_of(Gitlab::Git::Repository).to receive(:branch_exists?).and_return(false)
+ allow_any_instance_of(Repository).to receive(:fetch_source_branch!).and_return(true)
+ allow_any_instance_of(Gitlab::Git::Repository).to receive(:branch_exists?).and_return(false)
- expect_any_instance_of(Gitlab::Git::Repository).to receive(:create_branch).with('feature', 'DCBA')
- allow_any_instance_of(Gitlab::Git::Repository).to receive(:create_branch)
+ expect_any_instance_of(Gitlab::Git::Repository).to receive(:create_branch).with('feature', 'DCBA')
+ allow_any_instance_of(Gitlab::Git::Repository).to receive(:create_branch)
- project_tree_restorer = described_class.new(user: @user, shared: @shared, project: @project)
+ project_tree_restorer = described_class.new(user: @user, shared: @shared, project: @project)
- @restored_project_json = project_tree_restorer.restore
+ @restored_project_json = project_tree_restorer.restore
+ end
end
- end
- context 'JSON' do
- it 'restores models based on JSON' do
- expect(@restored_project_json).to be_truthy
+ after(:context) do
+ cleanup_artifacts_from_extract_archive('complex')
end
- it 'restore correct project features' do
- project = Project.find_by_path('project')
+ context 'JSON' do
+ it 'restores models based on JSON' do
+ expect(@restored_project_json).to be_truthy
+ end
- expect(project.project_feature.issues_access_level).to eq(ProjectFeature::PRIVATE)
- expect(project.project_feature.builds_access_level).to eq(ProjectFeature::PRIVATE)
- expect(project.project_feature.snippets_access_level).to eq(ProjectFeature::PRIVATE)
- expect(project.project_feature.wiki_access_level).to eq(ProjectFeature::PRIVATE)
- expect(project.project_feature.merge_requests_access_level).to eq(ProjectFeature::PRIVATE)
- end
+ it 'restore correct project features' do
+ project = Project.find_by_path('project')
- it 'has the project description' do
- expect(Project.find_by_path('project').description).to eq('Nisi et repellendus ut enim quo accusamus vel magnam.')
- end
+ expect(project.project_feature.issues_access_level).to eq(ProjectFeature::PRIVATE)
+ expect(project.project_feature.builds_access_level).to eq(ProjectFeature::PRIVATE)
+ expect(project.project_feature.snippets_access_level).to eq(ProjectFeature::PRIVATE)
+ expect(project.project_feature.wiki_access_level).to eq(ProjectFeature::PRIVATE)
+ expect(project.project_feature.merge_requests_access_level).to eq(ProjectFeature::PRIVATE)
+ end
- it 'has the same label associated to two issues' do
- expect(ProjectLabel.find_by_title('test2').issues.count).to eq(2)
- end
+ it 'has the project description' do
+ expect(Project.find_by_path('project').description).to eq('Nisi et repellendus ut enim quo accusamus vel magnam.')
+ end
- it 'has milestones associated to two separate issues' do
- expect(Milestone.find_by_description('test milestone').issues.count).to eq(2)
- end
+ it 'has the same label associated to two issues' do
+ expect(ProjectLabel.find_by_title('test2').issues.count).to eq(2)
+ end
+
+ it 'has milestones associated to two separate issues' do
+ expect(Milestone.find_by_description('test milestone').issues.count).to eq(2)
+ end
- context 'when importing a project with cached_markdown_version and note_html' do
- context 'for an Issue' do
- it 'does not import note_html' do
- note_content = 'Quo reprehenderit aliquam qui dicta impedit cupiditate eligendi'
- issue_note = Issue.find_by(description: 'Aliquam enim illo et possimus.').notes.select { |n| n.note.match(/#{note_content}/)}.first
+ context 'when importing a project with cached_markdown_version and note_html' do
+ context 'for an Issue' do
+ it 'does not import note_html' do
+ note_content = 'Quo reprehenderit aliquam qui dicta impedit cupiditate eligendi'
+ issue_note = Issue.find_by(description: 'Aliquam enim illo et possimus.').notes.select { |n| n.note.match(/#{note_content}/)}.first
- expect(issue_note.note_html).to match(/#{note_content}/)
+ expect(issue_note.note_html).to match(/#{note_content}/)
+ end
end
- end
- context 'for a Merge Request' do
- it 'does not import note_html' do
- note_content = 'Sit voluptatibus eveniet architecto quidem'
- merge_request_note = match_mr1_note(note_content)
+ context 'for a Merge Request' do
+ it 'does not import note_html' do
+ note_content = 'Sit voluptatibus eveniet architecto quidem'
+ merge_request_note = match_mr1_note(note_content)
- expect(merge_request_note.note_html).to match(/#{note_content}/)
+ expect(merge_request_note.note_html).to match(/#{note_content}/)
+ end
end
context 'merge request system note metadata' do
@@ -103,33 +110,32 @@ describe Gitlab::ImportExport::Project::TreeRestorer do
end
end
end
- end
- it 'creates a valid pipeline note' do
- expect(Ci::Pipeline.find_by_sha('sha-notes').notes).not_to be_empty
- end
+ it 'creates a valid pipeline note' do
+ expect(Ci::Pipeline.find_by_sha('sha-notes').notes).not_to be_empty
+ end
- it 'pipeline has the correct user ID' do
- expect(Ci::Pipeline.find_by_sha('sha-notes').user_id).to eq(@user.id)
- end
+ it 'pipeline has the correct user ID' do
+ expect(Ci::Pipeline.find_by_sha('sha-notes').user_id).to eq(@user.id)
+ end
- it 'restores pipelines with missing ref' do
- expect(Ci::Pipeline.where(ref: nil)).not_to be_empty
- end
+ it 'restores pipelines with missing ref' do
+ expect(Ci::Pipeline.where(ref: nil)).not_to be_empty
+ end
- it 'restores pipeline for merge request' do
- pipeline = Ci::Pipeline.find_by_sha('048721d90c449b244b7b4c53a9186b04330174ec')
+ it 'restores pipeline for merge request' do
+ pipeline = Ci::Pipeline.find_by_sha('048721d90c449b244b7b4c53a9186b04330174ec')
- expect(pipeline).to be_valid
- expect(pipeline.tag).to be_falsey
- expect(pipeline.source).to eq('merge_request_event')
- expect(pipeline.merge_request.id).to be > 0
- expect(pipeline.merge_request.target_branch).to eq('feature')
- expect(pipeline.merge_request.source_branch).to eq('feature_conflict')
- end
+ expect(pipeline).to be_valid
+ expect(pipeline.tag).to be_falsey
+ expect(pipeline.source).to eq('merge_request_event')
+ expect(pipeline.merge_request.id).to be > 0
+ expect(pipeline.merge_request.target_branch).to eq('feature')
+ expect(pipeline.merge_request.source_branch).to eq('feature_conflict')
+ end
- it 'restores pipelines based on ascending id order' do
- expected_ordered_shas = %w[
+ it 'restores pipelines based on ascending id order' do
+ expected_ordered_shas = %w[
2ea1f3dec713d940208fb5ce4a38765ecb5d3f73
ce84140e8b878ce6e7c4d298c7202ff38170e3ac
048721d90c449b244b7b4c53a9186b04330174ec
@@ -137,732 +143,749 @@ describe Gitlab::ImportExport::Project::TreeRestorer do
5f923865dde3436854e9ceb9cdb7815618d4e849
d2d430676773caa88cdaf7c55944073b2fd5561a
2ea1f3dec713d940208fb5ce4a38765ecb5d3f73
- ]
+ ]
- project = Project.find_by_path('project')
+ project = Project.find_by_path('project')
- project.ci_pipelines.order(:id).each_with_index do |pipeline, i|
- expect(pipeline['sha']).to eq expected_ordered_shas[i]
+ project.ci_pipelines.order(:id).each_with_index do |pipeline, i|
+ expect(pipeline['sha']).to eq expected_ordered_shas[i]
+ end
end
- end
- it 'preserves updated_at on issues' do
- issue = Issue.find_by(description: 'Aliquam enim illo et possimus.')
+ it 'preserves updated_at on issues' do
+ issue = Issue.find_by(description: 'Aliquam enim illo et possimus.')
- expect(issue.reload.updated_at.to_s).to eq('2016-06-14 15:02:47 UTC')
- end
+ expect(issue.reload.updated_at.to_s).to eq('2016-06-14 15:02:47 UTC')
+ end
- it 'has multiple issue assignees' do
- expect(Issue.find_by(title: 'Voluptatem').assignees).to contain_exactly(@user, *@existing_members)
- expect(Issue.find_by(title: 'Issue without assignees').assignees).to be_empty
- end
+ it 'has multiple issue assignees' do
+ expect(Issue.find_by(title: 'Voluptatem').assignees).to contain_exactly(@user, *@existing_members)
+ expect(Issue.find_by(title: 'Issue without assignees').assignees).to be_empty
+ end
- it 'restores timelogs for issues' do
- timelog = Issue.find_by(title: 'issue_with_timelogs').timelogs.last
+ it 'restores timelogs for issues' do
+ timelog = Issue.find_by(title: 'issue_with_timelogs').timelogs.last
- aggregate_failures do
- expect(timelog.time_spent).to eq(72000)
- expect(timelog.spent_at).to eq("2019-12-27T00:00:00.000Z")
+ aggregate_failures do
+ expect(timelog.time_spent).to eq(72000)
+ expect(timelog.spent_at).to eq("2019-12-27T00:00:00.000Z")
+ end
end
- end
- it 'contains the merge access levels on a protected branch' do
- expect(ProtectedBranch.first.merge_access_levels).not_to be_empty
- end
+ it 'contains the merge access levels on a protected branch' do
+ expect(ProtectedBranch.first.merge_access_levels).not_to be_empty
+ end
- it 'contains the push access levels on a protected branch' do
- expect(ProtectedBranch.first.push_access_levels).not_to be_empty
- end
+ it 'contains the push access levels on a protected branch' do
+ expect(ProtectedBranch.first.push_access_levels).not_to be_empty
+ end
- it 'contains the create access levels on a protected tag' do
- expect(ProtectedTag.first.create_access_levels).not_to be_empty
- end
+ it 'contains the create access levels on a protected tag' do
+ expect(ProtectedTag.first.create_access_levels).not_to be_empty
+ end
- it 'restores issue resource label events' do
- expect(Issue.find_by(title: 'Voluptatem').resource_label_events).not_to be_empty
- end
+ it 'restores issue resource label events' do
+ expect(Issue.find_by(title: 'Voluptatem').resource_label_events).not_to be_empty
+ end
- it 'restores merge requests resource label events' do
- expect(MergeRequest.find_by(title: 'MR1').resource_label_events).not_to be_empty
- end
+ it 'restores merge requests resource label events' do
+ expect(MergeRequest.find_by(title: 'MR1').resource_label_events).not_to be_empty
+ end
- it 'restores suggestion' do
- note = Note.find_by("note LIKE 'Saepe asperiores exercitationem non dignissimos laborum reiciendis et ipsum%'")
+ it 'restores suggestion' do
+ note = Note.find_by("note LIKE 'Saepe asperiores exercitationem non dignissimos laborum reiciendis et ipsum%'")
- expect(note.suggestions.count).to eq(1)
- expect(note.suggestions.first.from_content).to eq("Original line\n")
- end
+ expect(note.suggestions.count).to eq(1)
+ expect(note.suggestions.first.from_content).to eq("Original line\n")
+ end
- context 'event at forth level of the tree' do
- let(:event) { Event.find_by(action: 6) }
+ context 'event at forth level of the tree' do
+ let(:event) { Event.find_by(action: 6) }
- it 'restores the event' do
- expect(event).not_to be_nil
- end
+ it 'restores the event' do
+ expect(event).not_to be_nil
+ end
- it 'has the action' do
- expect(event.action).not_to be_nil
- end
+ it 'has the action' do
+ expect(event.action).not_to be_nil
+ end
- it 'event belongs to note, belongs to merge request, belongs to a project' do
- expect(event.note.noteable.project).not_to be_nil
+ it 'event belongs to note, belongs to merge request, belongs to a project' do
+ expect(event.note.noteable.project).not_to be_nil
+ end
end
- end
- it 'has the correct data for merge request diff files' do
- expect(MergeRequestDiffFile.where.not(diff: nil).count).to eq(55)
- end
+ it 'has the correct data for merge request diff files' do
+ expect(MergeRequestDiffFile.where.not(diff: nil).count).to eq(55)
+ end
- it 'has the correct data for merge request diff commits' do
- expect(MergeRequestDiffCommit.count).to eq(77)
- end
+ it 'has the correct data for merge request diff commits' do
+ expect(MergeRequestDiffCommit.count).to eq(77)
+ end
- it 'has the correct data for merge request latest_merge_request_diff' do
- MergeRequest.find_each do |merge_request|
- expect(merge_request.latest_merge_request_diff_id).to eq(merge_request.merge_request_diffs.maximum(:id))
+ it 'has the correct data for merge request latest_merge_request_diff' do
+ MergeRequest.find_each do |merge_request|
+ expect(merge_request.latest_merge_request_diff_id).to eq(merge_request.merge_request_diffs.maximum(:id))
+ end
end
- end
- it 'has labels associated to label links, associated to issues' do
- expect(Label.first.label_links.first.target).not_to be_nil
- end
+ it 'has labels associated to label links, associated to issues' do
+ expect(Label.first.label_links.first.target).not_to be_nil
+ end
- it 'has project labels' do
- expect(ProjectLabel.count).to eq(3)
- end
+ it 'has project labels' do
+ expect(ProjectLabel.count).to eq(3)
+ end
- it 'has no group labels' do
- expect(GroupLabel.count).to eq(0)
- end
+ it 'has no group labels' do
+ expect(GroupLabel.count).to eq(0)
+ end
- it 'has issue boards' do
- expect(Project.find_by_path('project').boards.count).to eq(1)
- end
+ it 'has issue boards' do
+ expect(Project.find_by_path('project').boards.count).to eq(1)
+ end
- it 'has lists associated with the issue board' do
- expect(Project.find_by_path('project').boards.find_by_name('TestBoardABC').lists.count).to eq(3)
- end
+ it 'has lists associated with the issue board' do
+ expect(Project.find_by_path('project').boards.find_by_name('TestBoardABC').lists.count).to eq(3)
+ end
- it 'has a project feature' do
- expect(@project.project_feature).not_to be_nil
- end
+ it 'has a project feature' do
+ expect(@project.project_feature).not_to be_nil
+ end
- it 'has custom attributes' do
- expect(@project.custom_attributes.count).to eq(2)
- end
+ it 'has custom attributes' do
+ expect(@project.custom_attributes.count).to eq(2)
+ end
- it 'has badges' do
- expect(@project.project_badges.count).to eq(2)
- end
+ it 'has badges' do
+ expect(@project.project_badges.count).to eq(2)
+ end
- it 'has snippets' do
- expect(@project.snippets.count).to eq(1)
- end
+ it 'has snippets' do
+ expect(@project.snippets.count).to eq(1)
+ end
- it 'has award emoji for a snippet' do
- award_emoji = @project.snippets.first.award_emoji
+ it 'has award emoji for a snippet' do
+ award_emoji = @project.snippets.first.award_emoji
- expect(award_emoji.map(&:name)).to contain_exactly('thumbsup', 'coffee')
- end
+ expect(award_emoji.map(&:name)).to contain_exactly('thumbsup', 'coffee')
+ end
- it 'snippet has notes' do
- expect(@project.snippets.first.notes.count).to eq(1)
- end
+ it 'snippet has notes' do
+ expect(@project.snippets.first.notes.count).to eq(1)
+ end
- it 'snippet has award emojis on notes' do
- award_emoji = @project.snippets.first.notes.first.award_emoji.first
+ it 'snippet has award emojis on notes' do
+ award_emoji = @project.snippets.first.notes.first.award_emoji.first
- expect(award_emoji.name).to eq('thumbsup')
- end
+ expect(award_emoji.name).to eq('thumbsup')
+ end
- it 'restores `ci_cd_settings` : `group_runners_enabled` setting' do
- expect(@project.ci_cd_settings.group_runners_enabled?).to eq(false)
- end
+ it 'restores `ci_cd_settings` : `group_runners_enabled` setting' do
+ expect(@project.ci_cd_settings.group_runners_enabled?).to eq(false)
+ end
- it 'restores `auto_devops`' do
- expect(@project.auto_devops_enabled?).to eq(true)
- expect(@project.auto_devops.deploy_strategy).to eq('continuous')
- end
+ it 'restores `auto_devops`' do
+ expect(@project.auto_devops_enabled?).to eq(true)
+ expect(@project.auto_devops.deploy_strategy).to eq('continuous')
+ end
- it 'restores the correct service' do
- expect(CustomIssueTrackerService.first).not_to be_nil
- end
+ it 'restores the correct service' do
+ expect(CustomIssueTrackerService.first).not_to be_nil
+ end
- it 'restores zoom meetings' do
- meetings = @project.issues.first.zoom_meetings
+ it 'restores zoom meetings' do
+ meetings = @project.issues.first.zoom_meetings
- expect(meetings.count).to eq(1)
- expect(meetings.first.url).to eq('https://zoom.us/j/123456789')
- end
+ expect(meetings.count).to eq(1)
+ expect(meetings.first.url).to eq('https://zoom.us/j/123456789')
+ end
- it 'restores sentry issues' do
- sentry_issue = @project.issues.first.sentry_issue
+ it 'restores sentry issues' do
+ sentry_issue = @project.issues.first.sentry_issue
- expect(sentry_issue.sentry_issue_identifier).to eq(1234567891)
- end
+ expect(sentry_issue.sentry_issue_identifier).to eq(1234567891)
+ end
- it 'has award emoji for an issue' do
- award_emoji = @project.issues.first.award_emoji.first
+ it 'has award emoji for an issue' do
+ award_emoji = @project.issues.first.award_emoji.first
- expect(award_emoji.name).to eq('musical_keyboard')
- end
+ expect(award_emoji.name).to eq('musical_keyboard')
+ end
- it 'has award emoji for a note in an issue' do
- award_emoji = @project.issues.first.notes.first.award_emoji.first
+ it 'has award emoji for a note in an issue' do
+ award_emoji = @project.issues.first.notes.first.award_emoji.first
- expect(award_emoji.name).to eq('clapper')
- end
+ expect(award_emoji.name).to eq('clapper')
+ end
- it 'restores container_expiration_policy' do
- policy = Project.find_by_path('project').container_expiration_policy
+ it 'restores container_expiration_policy' do
+ policy = Project.find_by_path('project').container_expiration_policy
- aggregate_failures do
- expect(policy).to be_an_instance_of(ContainerExpirationPolicy)
- expect(policy).to be_persisted
- expect(policy.cadence).to eq('3month')
+ aggregate_failures do
+ expect(policy).to be_an_instance_of(ContainerExpirationPolicy)
+ expect(policy).to be_persisted
+ expect(policy.cadence).to eq('3month')
+ end
end
- end
- it 'restores error_tracking_setting' do
- setting = @project.error_tracking_setting
+ it 'restores error_tracking_setting' do
+ setting = @project.error_tracking_setting
- aggregate_failures do
- expect(setting.api_url).to eq("https://gitlab.example.com/api/0/projects/sentry-org/sentry-project")
- expect(setting.project_name).to eq("Sentry Project")
- expect(setting.organization_name).to eq("Sentry Org")
+ aggregate_failures do
+ expect(setting.api_url).to eq("https://gitlab.example.com/api/0/projects/sentry-org/sentry-project")
+ expect(setting.project_name).to eq("Sentry Project")
+ expect(setting.organization_name).to eq("Sentry Org")
+ end
end
- end
- it 'restores external pull requests' do
- external_pr = @project.external_pull_requests.last
+ it 'restores external pull requests' do
+ external_pr = @project.external_pull_requests.last
- aggregate_failures do
- expect(external_pr.pull_request_iid).to eq(4)
- expect(external_pr.source_branch).to eq("feature")
- expect(external_pr.target_branch).to eq("master")
- expect(external_pr.status).to eq("open")
+ aggregate_failures do
+ expect(external_pr.pull_request_iid).to eq(4)
+ expect(external_pr.source_branch).to eq("feature")
+ expect(external_pr.target_branch).to eq("master")
+ expect(external_pr.status).to eq("open")
+ end
end
- end
- it 'restores pipeline schedules' do
- pipeline_schedule = @project.pipeline_schedules.last
+ it 'restores pipeline schedules' do
+ pipeline_schedule = @project.pipeline_schedules.last
- aggregate_failures do
- expect(pipeline_schedule.description).to eq('Schedule Description')
- expect(pipeline_schedule.ref).to eq('master')
- expect(pipeline_schedule.cron).to eq('0 4 * * 0')
- expect(pipeline_schedule.cron_timezone).to eq('UTC')
- expect(pipeline_schedule.active).to eq(true)
+ aggregate_failures do
+ expect(pipeline_schedule.description).to eq('Schedule Description')
+ expect(pipeline_schedule.ref).to eq('master')
+ expect(pipeline_schedule.cron).to eq('0 4 * * 0')
+ expect(pipeline_schedule.cron_timezone).to eq('UTC')
+ expect(pipeline_schedule.active).to eq(true)
+ end
end
- end
- it 'restores releases with links' do
- release = @project.releases.last
- link = release.links.last
+ it 'restores releases with links' do
+ release = @project.releases.last
+ link = release.links.last
- aggregate_failures do
- expect(release.tag).to eq('release-1.1')
- expect(release.description).to eq('Some release notes')
- expect(release.name).to eq('release-1.1')
- expect(release.sha).to eq('901de3a8bd5573f4a049b1457d28bc1592ba6bf9')
- expect(release.released_at).to eq('2019-12-26T10:17:14.615Z')
+ aggregate_failures do
+ expect(release.tag).to eq('release-1.1')
+ expect(release.description).to eq('Some release notes')
+ expect(release.name).to eq('release-1.1')
+ expect(release.sha).to eq('901de3a8bd5573f4a049b1457d28bc1592ba6bf9')
+ expect(release.released_at).to eq('2019-12-26T10:17:14.615Z')
- expect(link.url).to eq('http://localhost/namespace6/project6/-/jobs/140463678/artifacts/download')
- expect(link.name).to eq('release-1.1.dmg')
+ expect(link.url).to eq('http://localhost/namespace6/project6/-/jobs/140463678/artifacts/download')
+ expect(link.name).to eq('release-1.1.dmg')
+ end
end
- end
- context 'Merge requests' do
- it 'always has the new project as a target' do
- expect(MergeRequest.find_by_title('MR1').target_project).to eq(@project)
- end
+ context 'Merge requests' do
+ it 'always has the new project as a target' do
+ expect(MergeRequest.find_by_title('MR1').target_project).to eq(@project)
+ end
- it 'has the same source project as originally if source/target are the same' do
- expect(MergeRequest.find_by_title('MR1').source_project).to eq(@project)
- end
+ it 'has the same source project as originally if source/target are the same' do
+ expect(MergeRequest.find_by_title('MR1').source_project).to eq(@project)
+ end
- it 'has the new project as target if source/target differ' do
- expect(MergeRequest.find_by_title('MR2').target_project).to eq(@project)
- end
+ it 'has the new project as target if source/target differ' do
+ expect(MergeRequest.find_by_title('MR2').target_project).to eq(@project)
+ end
- it 'has no source if source/target differ' do
- expect(MergeRequest.find_by_title('MR2').source_project_id).to be_nil
- end
+ it 'has no source if source/target differ' do
+ expect(MergeRequest.find_by_title('MR2').source_project_id).to be_nil
+ end
- it 'has award emoji' do
- award_emoji = MergeRequest.find_by_title('MR1').award_emoji
+ it 'has award emoji' do
+ award_emoji = MergeRequest.find_by_title('MR1').award_emoji
- expect(award_emoji.map(&:name)).to contain_exactly('thumbsup', 'drum')
- end
+ expect(award_emoji.map(&:name)).to contain_exactly('thumbsup', 'drum')
+ end
- context 'notes' do
- it 'has award emoji' do
- merge_request_note = match_mr1_note('Sit voluptatibus eveniet architecto quidem')
- award_emoji = merge_request_note.award_emoji.first
+ context 'notes' do
+ it 'has award emoji' do
+ merge_request_note = match_mr1_note('Sit voluptatibus eveniet architecto quidem')
+ award_emoji = merge_request_note.award_emoji.first
- expect(award_emoji.name).to eq('tada')
+ expect(award_emoji.name).to eq('tada')
+ end
end
end
- end
- context 'tokens are regenerated' do
- it 'has new CI trigger tokens' do
- expect(Ci::Trigger.where(token: %w[cdbfasdf44a5958c83654733449e585 33a66349b5ad01fc00174af87804e40]))
- .to be_empty
- end
+ context 'tokens are regenerated' do
+ it 'has new CI trigger tokens' do
+ expect(Ci::Trigger.where(token: %w[cdbfasdf44a5958c83654733449e585 33a66349b5ad01fc00174af87804e40]))
+ .to be_empty
+ end
- it 'has a new CI build token' do
- expect(Ci::Build.where(token: 'abcd')).to be_empty
+ it 'has a new CI build token' do
+ expect(Ci::Build.where(token: 'abcd')).to be_empty
+ end
end
- end
- context 'has restored the correct number of records' do
- it 'has the correct number of merge requests' do
- expect(@project.merge_requests.size).to eq(9)
- end
+ context 'has restored the correct number of records' do
+ it 'has the correct number of merge requests' do
+ expect(@project.merge_requests.size).to eq(9)
+ end
- it 'only restores valid triggers' do
- expect(@project.triggers.size).to eq(1)
- end
+ it 'only restores valid triggers' do
+ expect(@project.triggers.size).to eq(1)
+ end
- it 'has the correct number of pipelines and statuses' do
- expect(@project.ci_pipelines.size).to eq(7)
+ it 'has the correct number of pipelines and statuses' do
+ expect(@project.ci_pipelines.size).to eq(7)
- @project.ci_pipelines.order(:id).zip([2, 0, 2, 2, 2, 2, 0])
- .each do |(pipeline, expected_status_size)|
- expect(pipeline.statuses.size).to eq(expected_status_size)
+ @project.ci_pipelines.order(:id).zip([2, 0, 2, 2, 2, 2, 0])
+ .each do |(pipeline, expected_status_size)|
+ expect(pipeline.statuses.size).to eq(expected_status_size)
+ end
end
end
- end
- context 'when restoring hierarchy of pipeline, stages and jobs' do
- it 'restores pipelines' do
- expect(Ci::Pipeline.all.count).to be 7
- end
+ context 'when restoring hierarchy of pipeline, stages and jobs' do
+ it 'restores pipelines' do
+ expect(Ci::Pipeline.all.count).to be 7
+ end
- it 'restores pipeline stages' do
- expect(Ci::Stage.all.count).to be 6
- end
+ it 'restores pipeline stages' do
+ expect(Ci::Stage.all.count).to be 6
+ end
- it 'correctly restores association between stage and a pipeline' do
- expect(Ci::Stage.all).to all(have_attributes(pipeline_id: a_value > 0))
- end
+ it 'correctly restores association between stage and a pipeline' do
+ expect(Ci::Stage.all).to all(have_attributes(pipeline_id: a_value > 0))
+ end
- it 'restores statuses' do
- expect(CommitStatus.all.count).to be 10
- end
+ it 'restores statuses' do
+ expect(CommitStatus.all.count).to be 10
+ end
- it 'correctly restores association between a stage and a job' do
- expect(CommitStatus.all).to all(have_attributes(stage_id: a_value > 0))
- end
+ it 'correctly restores association between a stage and a job' do
+ expect(CommitStatus.all).to all(have_attributes(stage_id: a_value > 0))
+ end
- it 'correctly restores association between a pipeline and a job' do
- expect(CommitStatus.all).to all(have_attributes(pipeline_id: a_value > 0))
- end
+ it 'correctly restores association between a pipeline and a job' do
+ expect(CommitStatus.all).to all(have_attributes(pipeline_id: a_value > 0))
+ end
- it 'restores a Hash for CommitStatus options' do
- expect(CommitStatus.all.map(&:options).compact).to all(be_a(Hash))
- end
+ it 'restores a Hash for CommitStatus options' do
+ expect(CommitStatus.all.map(&:options).compact).to all(be_a(Hash))
+ end
- it 'restores external pull request for the restored pipeline' do
- pipeline_with_external_pr = @project.ci_pipelines.find_by(source: 'external_pull_request_event')
+ it 'restores external pull request for the restored pipeline' do
+ pipeline_with_external_pr = @project.ci_pipelines.find_by(source: 'external_pull_request_event')
- expect(pipeline_with_external_pr.external_pull_request).to be_persisted
- end
+ expect(pipeline_with_external_pr.external_pull_request).to be_persisted
+ end
- it 'has no import failures' do
- expect(@project.import_failures.size).to eq 0
+ it 'has no import failures' do
+ expect(@project.import_failures.size).to eq 0
+ end
end
end
end
- end
- shared_examples 'restores group correctly' do |**results|
- it 'has group label' do
- expect(project.group.labels.size).to eq(results.fetch(:labels, 0))
- expect(project.group.labels.where(type: "GroupLabel").where.not(project_id: nil).count).to eq(0)
- end
+ shared_examples 'restores group correctly' do |**results|
+ it 'has group label' do
+ expect(project.group.labels.size).to eq(results.fetch(:labels, 0))
+ expect(project.group.labels.where(type: "GroupLabel").where.not(project_id: nil).count).to eq(0)
+ end
- it 'has group milestone' do
- expect(project.group.milestones.size).to eq(results.fetch(:milestones, 0))
- end
+ it 'has group milestone' do
+ expect(project.group.milestones.size).to eq(results.fetch(:milestones, 0))
+ end
- it 'has the correct visibility level' do
- # INTERNAL in the `project.json`, group's is PRIVATE
- expect(project.visibility_level).to eq(Gitlab::VisibilityLevel::PRIVATE)
+ it 'has the correct visibility level' do
+ # INTERNAL in the `project.json`, group's is PRIVATE
+ expect(project.visibility_level).to eq(Gitlab::VisibilityLevel::PRIVATE)
+ end
end
- end
- context 'project.json file access check' do
- let(:user) { create(:user) }
- let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') }
- let(:project_tree_restorer) do
- described_class.new(user: user, shared: shared, project: project)
- end
- let(:restored_project_json) { project_tree_restorer.restore }
+ context 'project.json file access check' do
+ let(:user) { create(:user) }
+ let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') }
+ let(:project_tree_restorer) do
+ described_class.new(user: user, shared: shared, project: project)
+ end
+ let(:restored_project_json) { project_tree_restorer.restore }
- it 'does not read a symlink' do
- Dir.mktmpdir do |tmpdir|
- setup_symlink(tmpdir, 'project.json')
- allow(shared).to receive(:export_path).and_call_original
+ it 'does not read a symlink' do
+ Dir.mktmpdir do |tmpdir|
+ setup_symlink(tmpdir, 'project.json')
+ allow(shared).to receive(:export_path).and_call_original
- expect(project_tree_restorer.restore).to eq(false)
- expect(shared.errors).to include('Incorrect JSON format')
+ expect(project_tree_restorer.restore).to eq(false)
+ expect(shared.errors).to include('invalid import format')
+ end
end
end
- end
- context 'Light JSON' do
- let(:user) { create(:user) }
- let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') }
- let(:project_tree_restorer) { described_class.new(user: user, shared: shared, project: project) }
- let(:restored_project_json) { project_tree_restorer.restore }
+ context 'Light JSON' do
+ let(:user) { create(:user) }
+ let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') }
+ let(:project_tree_restorer) { described_class.new(user: user, shared: shared, project: project) }
+ let(:restored_project_json) { project_tree_restorer.restore }
- context 'with a simple project' do
- before do
- setup_import_export_config('light')
- expect(restored_project_json).to eq(true)
- end
+ context 'with a simple project' do
+ before do
+ setup_import_export_config('light')
+ setup_reader(reader)
+
+ expect(restored_project_json).to eq(true)
+ end
+
+ after do
+ cleanup_artifacts_from_extract_archive('light')
+ end
+
+ it 'issue system note metadata restored successfully' do
+ note_content = 'created merge request !1 to address this issue'
+ note = project.issues.first.notes.select { |n| n.note.match(/#{note_content}/)}.first
+
+ expect(note.noteable_type).to eq('Issue')
+ expect(note.system).to eq(true)
+ expect(note.system_note_metadata.action).to eq('merge')
+ expect(note.system_note_metadata.commit_count).to be_nil
+ end
+
+ context 'when there is an existing build with build token' do
+ before do
+ create(:ci_build, token: 'abcd')
+ end
- it_behaves_like 'restores project successfully',
- issues: 1,
- labels: 2,
- label_with_priorities: 'A project label',
- milestones: 1,
- first_issue_labels: 1,
- services: 1
-
- it 'issue system note metadata restored successfully' do
- note_content = 'created merge request !1 to address this issue'
- note = project.issues.first.notes.select { |n| n.note.match(/#{note_content}/)}.first
-
- expect(note.noteable_type).to eq('Issue')
- expect(note.system).to eq(true)
- expect(note.system_note_metadata.action).to eq('merge')
- expect(note.system_note_metadata.commit_count).to be_nil
+ it_behaves_like 'restores project successfully',
+ issues: 1,
+ labels: 2,
+ label_with_priorities: 'A project label',
+ milestones: 1,
+ first_issue_labels: 1,
+ services: 1
+ end
+
+ context 'when there is an existing build with build token' do
+ before do
+ create(:ci_build, token: 'abcd')
+ end
+
+ it_behaves_like 'restores project successfully',
+ issues: 1,
+ labels: 2,
+ label_with_priorities: 'A project label',
+ milestones: 1,
+ first_issue_labels: 1
+ end
end
- context 'when there is an existing build with build token' do
+ context 'multiple pipelines reference the same external pull request' do
before do
- create(:ci_build, token: 'abcd')
+ setup_import_export_config('multi_pipeline_ref_one_external_pr')
+ setup_reader(reader)
+
+ expect(restored_project_json).to eq(true)
+ end
+
+ after do
+ cleanup_artifacts_from_extract_archive('multi_pipeline_ref_one_external_pr')
end
it_behaves_like 'restores project successfully',
- issues: 1,
- labels: 2,
- label_with_priorities: 'A project label',
- milestones: 1,
- first_issue_labels: 1
+ issues: 0,
+ labels: 0,
+ milestones: 0,
+ ci_pipelines: 2,
+ external_pull_requests: 1,
+ import_failures: 0
+
+ it 'restores external pull request for the restored pipelines' do
+ external_pr = project.external_pull_requests.first
+
+ project.ci_pipelines.each do |pipeline_with_external_pr|
+ expect(pipeline_with_external_pr.external_pull_request).to be_persisted
+ expect(pipeline_with_external_pr.external_pull_request).to eq(external_pr)
+ end
+ end
end
- end
- context 'multiple pipelines reference the same external pull request' do
- before do
- setup_import_export_config('multi_pipeline_ref_one_external_pr')
- expect(restored_project_json).to eq(true)
- end
+ context 'when post import action throw non-retriable exception' do
+ let(:exception) { StandardError.new('post_import_error') }
+
+ before do
+ setup_import_export_config('light')
+ setup_reader(reader)
- it_behaves_like 'restores project successfully',
- issues: 0,
- labels: 0,
- milestones: 0,
- ci_pipelines: 2,
- external_pull_requests: 1,
- import_failures: 0
+ expect(project)
+ .to receive(:merge_requests)
+ .and_raise(exception)
+ end
- it 'restores external pull request for the restored pipelines' do
- external_pr = project.external_pull_requests.first
+ after do
+ cleanup_artifacts_from_extract_archive('light')
+ end
- project.ci_pipelines.each do |pipeline_with_external_pr|
- expect(pipeline_with_external_pr.external_pull_request).to be_persisted
- expect(pipeline_with_external_pr.external_pull_request).to eq(external_pr)
+ it 'report post import error' do
+ expect(restored_project_json).to eq(false)
+ expect(shared.errors).to include('post_import_error')
end
end
- end
- context 'when post import action throw non-retriable exception' do
- let(:exception) { StandardError.new('post_import_error') }
+ context 'when post import action throw retriable exception one time' do
+ let(:exception) { GRPC::DeadlineExceeded.new }
- before do
- setup_import_export_config('light')
- expect(project)
- .to receive(:merge_requests)
- .and_raise(exception)
- end
+ before do
+ setup_import_export_config('light')
+ setup_reader(reader)
- it 'report post import error' do
- expect(restored_project_json).to eq(false)
- expect(shared.errors).to include('post_import_error')
- end
- end
+ expect(project)
+ .to receive(:merge_requests)
+ .and_raise(exception)
+ expect(project)
+ .to receive(:merge_requests)
+ .and_call_original
+ expect(restored_project_json).to eq(true)
+ end
- context 'when post import action throw retriable exception one time' do
- let(:exception) { GRPC::DeadlineExceeded.new }
+ after do
+ cleanup_artifacts_from_extract_archive('light')
+ end
- before do
- setup_import_export_config('light')
- expect(project)
- .to receive(:merge_requests)
- .and_raise(exception)
- expect(project)
- .to receive(:merge_requests)
- .and_call_original
- expect(restored_project_json).to eq(true)
- end
+ it_behaves_like 'restores project successfully',
+ issues: 1,
+ labels: 2,
+ label_with_priorities: 'A project label',
+ milestones: 1,
+ first_issue_labels: 1,
+ services: 1,
+ import_failures: 1
- it_behaves_like 'restores project successfully',
- issues: 1,
- labels: 2,
- label_with_priorities: 'A project label',
- milestones: 1,
- first_issue_labels: 1,
- services: 1,
- import_failures: 1
-
- it 'records the failures in the database' do
- import_failure = ImportFailure.last
-
- expect(import_failure.project_id).to eq(project.id)
- expect(import_failure.relation_key).to be_nil
- expect(import_failure.relation_index).to be_nil
- expect(import_failure.exception_class).to eq('GRPC::DeadlineExceeded')
- expect(import_failure.exception_message).to be_present
- expect(import_failure.correlation_id_value).not_to be_empty
- expect(import_failure.created_at).to be_present
- end
- end
+ it 'records the failures in the database' do
+ import_failure = ImportFailure.last
- context 'when the project has overridden params in import data' do
- before do
- setup_import_export_config('light')
+ expect(import_failure.project_id).to eq(project.id)
+ expect(import_failure.relation_key).to be_nil
+ expect(import_failure.relation_index).to be_nil
+ expect(import_failure.exception_class).to eq('GRPC::DeadlineExceeded')
+ expect(import_failure.exception_message).to be_present
+ expect(import_failure.correlation_id_value).not_to be_empty
+ expect(import_failure.created_at).to be_present
+ end
end
- it 'handles string versions of visibility_level' do
- # Project needs to be in a group for visibility level comparison
- # to happen
- group = create(:group)
- project.group = group
+ context 'when the project has overridden params in import data' do
+ before do
+ setup_import_export_config('light')
+ setup_reader(reader)
+ end
- project.create_import_data(data: { override_params: { visibility_level: Gitlab::VisibilityLevel::INTERNAL.to_s } })
+ after do
+ cleanup_artifacts_from_extract_archive('light')
+ end
- expect(restored_project_json).to eq(true)
- expect(project.visibility_level).to eq(Gitlab::VisibilityLevel::INTERNAL)
- end
+ it 'handles string versions of visibility_level' do
+ # Project needs to be in a group for visibility level comparison
+ # to happen
+ group = create(:group)
+ project.group = group
- it 'overwrites the params stored in the JSON' do
- project.create_import_data(data: { override_params: { description: "Overridden" } })
+ project.create_import_data(data: { override_params: { visibility_level: Gitlab::VisibilityLevel::INTERNAL.to_s } })
- expect(restored_project_json).to eq(true)
- expect(project.description).to eq("Overridden")
- end
+ expect(restored_project_json).to eq(true)
+ expect(project.visibility_level).to eq(Gitlab::VisibilityLevel::INTERNAL)
+ end
- it 'does not allow setting params that are excluded from import_export settings' do
- project.create_import_data(data: { override_params: { lfs_enabled: true } })
+ it 'overwrites the params stored in the JSON' do
+ project.create_import_data(data: { override_params: { description: "Overridden" } })
- expect(restored_project_json).to eq(true)
- expect(project.lfs_enabled).to be_falsey
- end
+ expect(restored_project_json).to eq(true)
+ expect(project.description).to eq("Overridden")
+ end
- it 'overrides project feature access levels' do
- access_level_keys = project.project_feature.attributes.keys.select { |a| a =~ /_access_level/ }
+ it 'does not allow setting params that are excluded from import_export settings' do
+ project.create_import_data(data: { override_params: { lfs_enabled: true } })
- # `pages_access_level` is not included, since it is not available in the public API
- # and has a dependency on project's visibility level
- # see ProjectFeature model
- access_level_keys.delete('pages_access_level')
+ expect(restored_project_json).to eq(true)
+ expect(project.lfs_enabled).to be_falsey
+ end
+
+ it 'overrides project feature access levels' do
+ access_level_keys = project.project_feature.attributes.keys.select { |a| a =~ /_access_level/ }
+
+ # `pages_access_level` is not included, since it is not available in the public API
+ # and has a dependency on project's visibility level
+ # see ProjectFeature model
+ access_level_keys.delete('pages_access_level')
- disabled_access_levels = Hash[access_level_keys.collect { |item| [item, 'disabled'] }]
+ disabled_access_levels = Hash[access_level_keys.collect { |item| [item, 'disabled'] }]
- project.create_import_data(data: { override_params: disabled_access_levels })
+ project.create_import_data(data: { override_params: disabled_access_levels })
- expect(restored_project_json).to eq(true)
+ expect(restored_project_json).to eq(true)
- aggregate_failures do
- access_level_keys.each do |key|
- expect(project.public_send(key)).to eq(ProjectFeature::DISABLED)
+ aggregate_failures do
+ access_level_keys.each do |key|
+ expect(project.public_send(key)).to eq(ProjectFeature::DISABLED)
+ end
end
end
end
- end
- context 'with a project that has a group' do
- let!(:project) do
- create(:project,
- :builds_disabled,
- :issues_disabled,
- name: 'project',
- path: 'project',
- group: create(:group, visibility_level: Gitlab::VisibilityLevel::PRIVATE))
- end
+ context 'with a project that has a group' do
+ let!(:project) do
+ create(:project,
+ :builds_disabled,
+ :issues_disabled,
+ name: 'project',
+ path: 'project',
+ group: create(:group, visibility_level: Gitlab::VisibilityLevel::PRIVATE))
+ end
- before do
- setup_import_export_config('group')
- expect(restored_project_json).to eq(true)
- end
+ before do
+ setup_import_export_config('group')
+ setup_reader(reader)
- it_behaves_like 'restores project successfully',
- issues: 3,
- labels: 2,
- label_with_priorities: 'A project label',
- milestones: 2,
- first_issue_labels: 1
-
- it_behaves_like 'restores group correctly',
- labels: 0,
- milestones: 0,
- first_issue_labels: 1
-
- it 'restores issue states' do
- expect(project.issues.with_state(:closed).count).to eq(1)
- expect(project.issues.with_state(:opened).count).to eq(2)
- end
- end
+ expect(restored_project_json).to eq(true)
+ end
- context 'with existing group models' do
- let!(:project) do
- create(:project,
- :builds_disabled,
- :issues_disabled,
- name: 'project',
- path: 'project',
- group: create(:group))
- end
+ after do
+ cleanup_artifacts_from_extract_archive('group')
+ end
- before do
- setup_import_export_config('light')
- end
+ it_behaves_like 'restores project successfully',
+ issues: 3,
+ labels: 2,
+ label_with_priorities: 'A project label',
+ milestones: 2,
+ first_issue_labels: 1
- it 'does not import any templated services' do
- expect(restored_project_json).to eq(true)
+ it_behaves_like 'restores group correctly',
+ labels: 0,
+ milestones: 0,
+ first_issue_labels: 1
- expect(project.services.where(template: true).count).to eq(0)
+ it 'restores issue states' do
+ expect(project.issues.with_state(:closed).count).to eq(1)
+ expect(project.issues.with_state(:opened).count).to eq(2)
+ end
end
- it 'does not import any instance services' do
- expect(restored_project_json).to eq(true)
+ context 'with existing group models' do
+ let!(:project) do
+ create(:project,
+ :builds_disabled,
+ :issues_disabled,
+ name: 'project',
+ path: 'project',
+ group: create(:group))
+ end
- expect(project.services.where(instance: true).count).to eq(0)
- end
+ before do
+ setup_import_export_config('light')
+ setup_reader(reader)
+ end
- it 'imports labels' do
- create(:group_label, name: 'Another label', group: project.group)
+ after do
+ cleanup_artifacts_from_extract_archive('light')
+ end
- expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error)
+ it 'does not import any templated services' do
+ expect(restored_project_json).to eq(true)
- expect(restored_project_json).to eq(true)
- expect(project.labels.count).to eq(1)
- end
+ expect(project.services.where(template: true).count).to eq(0)
+ end
- it 'imports milestones' do
- create(:milestone, name: 'A milestone', group: project.group)
+ it 'does not import any instance services' do
+ expect(restored_project_json).to eq(true)
- expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error)
+ expect(project.services.where(instance: true).count).to eq(0)
+ end
- expect(restored_project_json).to eq(true)
- expect(project.group.milestones.count).to eq(1)
- expect(project.milestones.count).to eq(0)
- end
- end
+ it 'imports labels' do
+ create(:group_label, name: 'Another label', group: project.group)
- context 'with clashing milestones on IID' do
- let!(:project) do
- create(:project,
- :builds_disabled,
- :issues_disabled,
- name: 'project',
- path: 'project',
- group: create(:group))
- end
+ expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error)
- before do
- setup_import_export_config('milestone-iid')
- end
+ expect(restored_project_json).to eq(true)
+ expect(project.labels.count).to eq(1)
+ end
- it 'preserves the project milestone IID' do
- expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error)
+ it 'imports milestones' do
+ create(:milestone, name: 'A milestone', group: project.group)
- expect(restored_project_json).to eq(true)
- expect(project.milestones.count).to eq(2)
- expect(Milestone.find_by_title('Another milestone').iid).to eq(1)
- expect(Milestone.find_by_title('Group-level milestone').iid).to eq(2)
- end
- end
+ expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error)
- context 'with external authorization classification labels' do
- before do
- setup_import_export_config('light')
+ expect(restored_project_json).to eq(true)
+ expect(project.group.milestones.count).to eq(1)
+ expect(project.milestones.count).to eq(0)
+ end
end
- it 'converts empty external classification authorization labels to nil' do
- project.create_import_data(data: { override_params: { external_authorization_classification_label: "" } })
+ context 'with clashing milestones on IID' do
+ let!(:project) do
+ create(:project,
+ :builds_disabled,
+ :issues_disabled,
+ name: 'project',
+ path: 'project',
+ group: create(:group))
+ end
- expect(restored_project_json).to eq(true)
- expect(project.external_authorization_classification_label).to be_nil
- end
+ before do
+ setup_import_export_config('milestone-iid')
+ setup_reader(reader)
+ end
- it 'preserves valid external classification authorization labels' do
- project.create_import_data(data: { override_params: { external_authorization_classification_label: "foobar" } })
+ after do
+ cleanup_artifacts_from_extract_archive('milestone-iid')
+ end
- expect(restored_project_json).to eq(true)
- expect(project.external_authorization_classification_label).to eq("foobar")
- end
- end
- end
+ it 'preserves the project milestone IID' do
+ expect_any_instance_of(Gitlab::ImportExport::Shared).not_to receive(:error)
- context 'Minimal JSON' do
- let(:project) { create(:project) }
- let(:user) { create(:user) }
- let(:tree_hash) { { 'visibility_level' => visibility } }
- let(:restorer) do
- described_class.new(user: user, shared: shared, project: project)
- end
+ expect(restored_project_json).to eq(true)
+ expect(project.milestones.count).to eq(2)
+ expect(Milestone.find_by_title('Another milestone').iid).to eq(1)
+ expect(Milestone.find_by_title('Group-level milestone').iid).to eq(2)
+ end
+ end
- before do
- allow_any_instance_of(Gitlab::ImportExport::JSON::LegacyReader::File).to receive(:valid?).and_return(true)
- allow_any_instance_of(Gitlab::ImportExport::JSON::LegacyReader::File).to receive(:tree_hash) { tree_hash }
- end
+ context 'with external authorization classification labels' do
+ before do
+ setup_import_export_config('light')
+ setup_reader(reader)
+ end
- context 'no group visibility' do
- let(:visibility) { Gitlab::VisibilityLevel::PRIVATE }
+ after do
+ cleanup_artifacts_from_extract_archive('light')
+ end
- it 'uses the project visibility' do
- expect(restorer.restore).to eq(true)
- expect(restorer.project.visibility_level).to eq(visibility)
- end
- end
+ it 'converts empty external classification authorization labels to nil' do
+ project.create_import_data(data: { override_params: { external_authorization_classification_label: "" } })
- context 'with restricted internal visibility' do
- describe 'internal project' do
- let(:visibility) { Gitlab::VisibilityLevel::INTERNAL }
+ expect(restored_project_json).to eq(true)
+ expect(project.external_authorization_classification_label).to be_nil
+ end
- it 'uses private visibility' do
- stub_application_setting(restricted_visibility_levels: [Gitlab::VisibilityLevel::INTERNAL])
+ it 'preserves valid external classification authorization labels' do
+ project.create_import_data(data: { override_params: { external_authorization_classification_label: "foobar" } })
- expect(restorer.restore).to eq(true)
- expect(restorer.project.visibility_level).to eq(Gitlab::VisibilityLevel::PRIVATE)
+ expect(restored_project_json).to eq(true)
+ expect(project.external_authorization_classification_label).to eq("foobar")
end
end
end
- context 'with group visibility' do
- before do
- group = create(:group, visibility_level: group_visibility)
-
- project.update(group: group)
+ context 'Minimal JSON' do
+ let(:project) { create(:project) }
+ let(:user) { create(:user) }
+ let(:tree_hash) { { 'visibility_level' => visibility } }
+ let(:restorer) do
+ described_class.new(user: user, shared: shared, project: project)
end
- context 'private group visibility' do
- let(:group_visibility) { Gitlab::VisibilityLevel::PRIVATE }
- let(:visibility) { Gitlab::VisibilityLevel::PUBLIC }
-
- it 'uses the group visibility' do
- expect(restorer.restore).to eq(true)
- expect(restorer.project.visibility_level).to eq(group_visibility)
- end
+ before do
+ allow_any_instance_of(Gitlab::ImportExport::JSON::LegacyReader::File).to receive(:exist?).and_return(true)
+ allow_any_instance_of(Gitlab::ImportExport::JSON::NdjsonReader).to receive(:exist?).and_return(false)
+ allow_any_instance_of(Gitlab::ImportExport::JSON::LegacyReader::File).to receive(:tree_hash) { tree_hash }
end
- context 'public group visibility' do
- let(:group_visibility) { Gitlab::VisibilityLevel::PUBLIC }
+ context 'no group visibility' do
let(:visibility) { Gitlab::VisibilityLevel::PRIVATE }
it 'uses the project visibility' do
@@ -871,17 +894,11 @@ describe Gitlab::ImportExport::Project::TreeRestorer do
end
end
- context 'internal group visibility' do
- let(:group_visibility) { Gitlab::VisibilityLevel::INTERNAL }
- let(:visibility) { Gitlab::VisibilityLevel::PUBLIC }
-
- it 'uses the group visibility' do
- expect(restorer.restore).to eq(true)
- expect(restorer.project.visibility_level).to eq(group_visibility)
- end
+ context 'with restricted internal visibility' do
+ describe 'internal project' do
+ let(:visibility) { Gitlab::VisibilityLevel::INTERNAL }
- context 'with restricted internal visibility' do
- it 'sets private visibility' do
+ it 'uses private visibility' do
stub_application_setting(restricted_visibility_levels: [Gitlab::VisibilityLevel::INTERNAL])
expect(restorer.restore).to eq(true)
@@ -889,43 +906,116 @@ describe Gitlab::ImportExport::Project::TreeRestorer do
end
end
end
- end
- end
- context 'JSON with invalid records' do
- subject(:restored_project_json) { project_tree_restorer.restore }
+ context 'with group visibility' do
+ before do
+ group = create(:group, visibility_level: group_visibility)
+
+ project.update(group: group)
+ end
- let(:user) { create(:user) }
- let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') }
- let(:project_tree_restorer) { described_class.new(user: user, shared: shared, project: project) }
+ context 'private group visibility' do
+ let(:group_visibility) { Gitlab::VisibilityLevel::PRIVATE }
+ let(:visibility) { Gitlab::VisibilityLevel::PUBLIC }
- before do
- setup_import_export_config('with_invalid_records')
+ it 'uses the group visibility' do
+ expect(restorer.restore).to eq(true)
+ expect(restorer.project.visibility_level).to eq(group_visibility)
+ end
+ end
+
+ context 'public group visibility' do
+ let(:group_visibility) { Gitlab::VisibilityLevel::PUBLIC }
+ let(:visibility) { Gitlab::VisibilityLevel::PRIVATE }
+
+ it 'uses the project visibility' do
+ expect(restorer.restore).to eq(true)
+ expect(restorer.project.visibility_level).to eq(visibility)
+ end
+ end
- subject
+ context 'internal group visibility' do
+ let(:group_visibility) { Gitlab::VisibilityLevel::INTERNAL }
+ let(:visibility) { Gitlab::VisibilityLevel::PUBLIC }
+
+ it 'uses the group visibility' do
+ expect(restorer.restore).to eq(true)
+ expect(restorer.project.visibility_level).to eq(group_visibility)
+ end
+
+ context 'with restricted internal visibility' do
+ it 'sets private visibility' do
+ stub_application_setting(restricted_visibility_levels: [Gitlab::VisibilityLevel::INTERNAL])
+
+ expect(restorer.restore).to eq(true)
+ expect(restorer.project.visibility_level).to eq(Gitlab::VisibilityLevel::PRIVATE)
+ end
+ end
+ end
+ end
end
- context 'when failures occur because a relation fails to be processed' do
- it_behaves_like 'restores project successfully',
- issues: 0,
- labels: 0,
- label_with_priorities: nil,
- milestones: 1,
- first_issue_labels: 0,
- services: 0,
- import_failures: 1
-
- it 'records the failures in the database' do
- import_failure = ImportFailure.last
-
- expect(import_failure.project_id).to eq(project.id)
- expect(import_failure.relation_key).to eq('milestones')
- expect(import_failure.relation_index).to be_present
- expect(import_failure.exception_class).to eq('ActiveRecord::RecordInvalid')
- expect(import_failure.exception_message).to be_present
- expect(import_failure.correlation_id_value).not_to be_empty
- expect(import_failure.created_at).to be_present
+ context 'JSON with invalid records' do
+ subject(:restored_project_json) { project_tree_restorer.restore }
+
+ let(:user) { create(:user) }
+ let!(:project) { create(:project, :builds_disabled, :issues_disabled, name: 'project', path: 'project') }
+ let(:project_tree_restorer) { described_class.new(user: user, shared: shared, project: project) }
+
+ before do
+ setup_import_export_config('with_invalid_records')
+ setup_reader(reader)
+
+ subject
+ end
+
+ after do
+ cleanup_artifacts_from_extract_archive('with_invalid_records')
end
+
+ context 'when failures occur because a relation fails to be processed' do
+ it_behaves_like 'restores project successfully',
+ issues: 0,
+ labels: 0,
+ label_with_priorities: nil,
+ milestones: 1,
+ first_issue_labels: 0,
+ services: 0,
+ import_failures: 1
+
+ it 'records the failures in the database' do
+ import_failure = ImportFailure.last
+
+ expect(import_failure.project_id).to eq(project.id)
+ expect(import_failure.relation_key).to eq('milestones')
+ expect(import_failure.relation_index).to be_present
+ expect(import_failure.exception_class).to eq('ActiveRecord::RecordInvalid')
+ expect(import_failure.exception_message).to be_present
+ expect(import_failure.correlation_id_value).not_to be_empty
+ expect(import_failure.created_at).to be_present
+ end
+ end
+ end
+ end
+
+ context 'enable ndjson import' do
+ before_all do
+ # Test suite `restore project tree` run `project_tree_restorer.restore` in `before_all`.
+ # `Enable all features by default for testing` happens in `before(:each)`
+ # So it requires manually enable feature flag to allow ndjson_reader
+ Feature.enable(:project_import_ndjson)
+ end
+
+ it_behaves_like 'project tree restorer work properly', :legacy_reader
+
+ it_behaves_like 'project tree restorer work properly', :ndjson_reader
+ end
+
+ context 'disable ndjson import' do
+ before do
+ stub_feature_flags(project_import_ndjson: false)
end
+
+ it_behaves_like 'project tree restorer work properly', :legacy_reader
end
end
diff --git a/spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb b/spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb
index 52e1efa70e0..0b58a75220d 100644
--- a/spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb
+++ b/spec/lib/gitlab/import_export/relation_tree_restorer_spec.rb
@@ -14,7 +14,7 @@ describe Gitlab::ImportExport::RelationTreeRestorer do
let(:user) { create(:user) }
let(:shared) { Gitlab::ImportExport::Shared.new(importable) }
- let(:attributes) { {} }
+ let(:attributes) { relation_reader.consume_attributes(importable_name) }
let(:members_mapper) do
Gitlab::ImportExport::MembersMapper.new(exported_members: {}, user: user, importable: importable)
@@ -30,7 +30,7 @@ describe Gitlab::ImportExport::RelationTreeRestorer do
relation_factory: relation_factory,
reader: reader,
importable: importable,
- importable_path: nil,
+ importable_path: importable_path,
importable_attributes: attributes
)
end
@@ -94,21 +94,24 @@ describe Gitlab::ImportExport::RelationTreeRestorer do
end
context 'when restoring a project' do
- let(:path) { 'spec/fixtures/lib/gitlab/import_export/complex/project.json' }
let(:importable) { create(:project, :builds_enabled, :issues_disabled, name: 'project', path: 'project') }
+ let(:importable_name) { 'project' }
+ let(:importable_path) { 'project' }
let(:object_builder) { Gitlab::ImportExport::Project::ObjectBuilder }
let(:relation_factory) { Gitlab::ImportExport::Project::RelationFactory }
let(:reader) { Gitlab::ImportExport::Reader.new(shared: shared) }
context 'using legacy reader' do
+ let(:path) { 'spec/fixtures/lib/gitlab/import_export/complex/project.json' }
let(:relation_reader) do
Gitlab::ImportExport::JSON::LegacyReader::File.new(
path,
- relation_names: reader.project_relation_names
+ relation_names: reader.project_relation_names,
+ allowed_path: 'project'
)
end
- let(:attributes) { relation_reader.consume_attributes(nil) }
+ let(:attributes) { relation_reader.consume_attributes('project') }
it_behaves_like 'import project successfully'
@@ -118,6 +121,21 @@ describe Gitlab::ImportExport::RelationTreeRestorer do
include_examples 'logging of relations creation'
end
+
+ context 'using ndjson reader' do
+ let(:path) { 'spec/fixtures/lib/gitlab/import_export/complex/tree' }
+ let(:relation_reader) { Gitlab::ImportExport::JSON::NdjsonReader.new(path) }
+
+ before :all do
+ extract_archive('spec/fixtures/lib/gitlab/import_export/complex', 'tree.tar.gz')
+ end
+
+ after :all do
+ cleanup_artifacts_from_extract_archive('complex')
+ end
+
+ it_behaves_like 'import project successfully'
+ end
end
end
@@ -125,9 +143,16 @@ describe Gitlab::ImportExport::RelationTreeRestorer do
let(:path) { 'spec/fixtures/lib/gitlab/import_export/group_exports/no_children/group.json' }
let(:group) { create(:group) }
let(:importable) { create(:group, parent: group) }
+ let(:importable_name) { nil }
+ let(:importable_path) { nil }
let(:object_builder) { Gitlab::ImportExport::Group::ObjectBuilder }
let(:relation_factory) { Gitlab::ImportExport::Group::RelationFactory }
- let(:relation_reader) { Gitlab::ImportExport::JSON::LegacyReader::File.new(path, relation_names: reader.group_relation_names) }
+ let(:relation_reader) do
+ Gitlab::ImportExport::JSON::LegacyReader::File.new(
+ path,
+ relation_names: reader.group_relation_names)
+ end
+
let(:reader) do
Gitlab::ImportExport::Reader.new(
shared: shared,
@@ -135,6 +160,10 @@ describe Gitlab::ImportExport::RelationTreeRestorer do
)
end
+ it 'restores group tree' do
+ expect(subject).to eq(true)
+ end
+
include_examples 'logging of relations creation'
end
end