Entity Framework Core – Isolation Of Integration Tests

When working with Entity Framework Core (EF) a lot of code can be tested using the In-Memory database provider but sometimes you want (or have) to go to the real database. For example, you are using not just LINQ but custom SQL statements due to performance reasons or you want to check that a specific exception is thrown by the database under some conditions like when having a primary key violation.

In diesem Artikel:

pg
Pawel Gerr ist Architekt und Consultant bei Thinktecture. Er hat sich auf .NET Core Backends spezialisiert und kennt Entity Framework von vorne bis hinten.
The biggest challenge of integration tests is the isolation of one test from another. In this post we will look at 3 options how to do that. The code for the 3rd option is on GitHub: PawelGerr/EntityFrameworkCore-Demos

Remarks: in my demos I’m using 3rd party libs: FluentAssertions and xunit.

Given is a DemoRepository with a method AddProduct that we want to test. (The code kept oversimplified for clarity reasons)

				
					public class DemoRepository
{
  ...

  public void AddProduct(Guid id)
  {
    _dbContext.Products.Add(new Product { Id = id });
    _dbContext.SaveChanges();
  }
}

				
			

Using Transaction Scopes

EF Core added support for [TransactionScope](https://docs.microsoft.com/en-us/ef/core/saving/transactions#using-systemtransactions) in version 2.1.

The isolation of tests via TransactionScope is very simple just wrap the call AddProduct into a TransactionScope to revert all changes at the end of the test. But, there are few preconditions. The testing method must not starting transactions using BeginTransaction() or it has to use a TransactionScope as well.

Also, I recommend to read my other blog post: Entity Framework Core: Use TransactionScope with Caution!

				
					public DemoRepositoryTests()
{
  _dbContext = CreateDbContext();
  _repository = new DemoRepository(_dbContext);
}

[Fact]
public void Should_add_new_product()
{
  var productId = new Guid("DBD9439E-6FFD-4719-93C7-3F7FA64D2220");

  using(var scope = new TransactionScope())
  {
    _repository.AddProduct(productId);

    _dbContext.Products.FirstOrDefault(p => p.Id == productId).Should().NotBeNull();

    // the transaction is going to be rolled back because the scope is not completed
    // scope.Complete();
  } 
}

				
			

Using new Databases

Creating a new database for each test is very easy but the tests are very time consuming. On my machine each test takes about 10 seconds to create and to delete a database on the fly.

The steps of each test are: generate a new database name, create the database by running EF migrations and delete the database in the end.

				
					public class DemoRepositoryTests : IDisposable
{
  private readonly DemoDbContext _dbContext;
  private readonly DemoRepository _repository;
  private readonly string _databaseName;

  public DemoRepositoryTests()
  {
    _databaseName = Guid.NewGuid().ToString();

    var options = new DbContextOptionsBuilder<DemoDbContext>()
              .UseSqlServer($"Server=(local);Database={_databaseName};...")
              .Options;

    _dbContext = new DemoDbContext(options);
    _dbContext.Database.Migrate();

    _repository = new DemoRepository(_dbContext);
  }

  // Tests come here

  public void Dispose()
  {
    _dbContext.Database.ExecuteSqlCommand((string)$"DROP DATABASE [{_databaseName}]");
  }
}
				
			

Using different Database Schemas

The 3rd option is to use the same database but different schemas. The creation of a new schema and running EF migrations usually takes less than 50 ms, which is totally acceptable for an integration test. The prerequisites to run queries with different schemas are schema-aware instances of DbContext and schema-aware EF migrations. Read my blog posts for more information about how to change the database schema at runtime:

The class executing integration tests consists of 2 parts: creation of the tables in constructor and the deletion of them in Dispose().

I’m using a generic base class to use the same logic for different types of DbContext.

In the constructor we generate the name of the schema using Guid.NewGuid(), create DbContextOptions using DbSchemaAwareMigrationAssembly and DbSchemaAwareModelCacheKeyFactory described in my previous posts, create the DbContext and run the EF migrations. The database is now fully prepared for executing tests. After execution of the tests the EF migrations are rolled back using IMigrator.Migrate("0"), the EF history table __EFMigrationsHistory is deleted and newly generated schema is dropped.

				
					public abstract class IntegrationTestsBase<T> : IDisposable
  where T : DbContext
{
  private readonly string _schema;
  private readonly string _historyTableName;
  private readonly DbContextOptions<T> _options;

  protected T DbContext { get; }

  protected IntegrationTestsBase()
  {
    _schema = Guid.NewGuid().ToString("N");
    _historyTableName = "__EFMigrationsHistory";

    _options = CreateOptions();
    DbContext = CreateContext();
    DbContext.Database.Migrate();
  }

  protected abstract T CreateContext(DbContextOptions<T> options, 
                                     IDbContextSchema schema);

  protected T CreateContext()
  {
    return CreateContext(_options, new DbContextSchema(_schema));
  }

  private DbContextOptions<T> CreateOptions()
  {
    return new DbContextOptionsBuilder<T>()
        .UseSqlServer($"Server=(local);Database=Demo;...", 
                builder => builder.MigrationsHistoryTable(_historyTableName, _schema))
        .ReplaceService<IMigrationsAssembly, DbSchemaAwareMigrationAssembly>()
        .ReplaceService<IModelCacheKeyFactory, DbSchemaAwareModelCacheKeyFactory>()
        .Options;
  }

  public void Dispose()
  {
    DbContext.GetService<IMigrator>().Migrate("0");
    DbContext.Database.ExecuteSqlCommand(
           (string)$"DROP TABLE [{_schema}].[{_historyTableName}]");
    DbContext.Database.ExecuteSqlCommand((string)$"DROP SCHEMA [{_schema}]");

    DbContext?.Dispose();
  }
}
				
			

The usage of the base class looks as follows

				
					public class DemoRepositoryTests : IntegrationTestsBase<DemoDbContext>
{
  private readonly DemoRepository _repository;

  public DemoRepositoryTests()
  {
    _repository = new DemoRepository(DbContext);
  }

  protected override DemoDbContext CreateContext(DbContextOptions<DemoDbContext> options, 
                                                 IDbContextSchema schema)
  {
    return new DemoDbContext(options, schema);
  }

  [Fact]
  public void Should_add_new_product()
  {
    var productId = new Guid("DBD9439E-6FFD-4719-93C7-3F7FA64D2220");

    _repository.AddProduct(productId);

    DbContext.Products.FirstOrDefault(p => p.Id == productId).Should().NotBeNull();
  }
}
				
			

Happy testing!

Kostenloser
Newsletter

Aktuelle Artikel, Screencasts, Webinare und Interviews unserer Experten für Sie

Verpassen Sie keine Inhalte zu Angular, .NET Core, Blazor, Azure und Kubernetes und melden Sie sich zu unserem kostenlosen monatlichen Dev-Newsletter an.

Newsletter Anmeldung
Diese Artikel könnten Sie interessieren
.NET
pg

Discriminated Unions in .NET: Modeling States and Variants

Domain models often involve concepts that exist in multiple distinct states or variations. Traditional approaches using enums and nullable properties can lead to invalid states and scattered logic. This article explores how discriminated unions provide a structured, type-safe way to model domain variants in .NET, aligning perfectly with Domain-Driven Design principles while enforcing invariants at the type level.
06.10.2025
.NET
pg

Discriminated Unions in .NET: Modeling States and Variants

Domain models often involve concepts that exist in multiple distinct states or variations. Traditional approaches using enums and nullable properties can lead to invalid states and scattered logic. This article explores how discriminated unions provide a structured, type-safe way to model domain variants in .NET, aligning perfectly with Domain-Driven Design principles while enforcing invariants at the type level.
06.10.2025
.NET
pg

Smart Enums in .NET: Integration with Frameworks and Libraries

Learn how to seamlessly integrate Smart Enums with essential .NET frameworks and libraries. This article covers practical solutions for JSON serialization, ASP.NET Core model binding for both Minimal APIs and MVC controllers, and Entity Framework Core persistence using value converters. Discover how Thinktecture.Runtime.Extensions provides dedicated packages to eliminate integration friction and maintain type safety across your application stack.
21.09.2025
.NET
pg

Value Objects in .NET: Enhancing Business Semantics

Value objects are fundamental building blocks in Domain-Driven Design, serving far more than simple data wrappers. This article explores their strategic importance in bridging technical code and business concepts, enforcing domain rules, and fostering clearer communication with domain experts. Learn how to build robust aggregates, cultivate ubiquitous language, and encapsulate domain-specific behavior using Thinktecture.Runtime.Extensions in .NET applications.
16.09.2025
.NET
pg

Pattern Matching with Discriminated Unions in .NET

Traditional C# pattern matching with switch statements and if/else chains is error-prone and doesn't guarantee exhaustive handling of all cases. When you add new types or states, it's easy to miss updating conditional logic, leading to runtime bugs. The library Thinktecture.Runtime.Extensions solves this with built-in Switch and Map methods for discriminated unions that enforce compile-time exhaustiveness checking.
26.08.2025
.NET
pg

Value Objects in .NET: Integration with Frameworks and Libraries

Value Objects in .NET provide a structured way to improve consistency and maintainability in domain modeling. This article examines their integration with popular frameworks and libraries, highlighting best practices for seamless implementation. From working with Entity Framework to leveraging their advantages in ASP.NET, we explore how Value Objects can be effectively incorporated into various architectures. By understanding their role in framework integration, developers can optimize data handling and enhance code clarity without unnecessary complexity.
12.08.2025