Type checking and conversion are essential operations in C#'s object-oriented programming model.
The is and as keywords provide elegant solutions for safely working with types at runtime. Understanding when and how to use each can significantly improve your code's robustness and readability.
is
as
The is operator evaluates whether an object is compatible with a given type, returning a boolean result.
object value = "Hello, World!"; // Check if value is a string if (value is string) { Console.WriteLine("value is a string"); }
// Type checking with declaration if (value is string message) { // message is now a string variable containing the value Console.WriteLine($"Length: {message.Length}"); }
// Check type and condition in one step if (value is string { Length: > 5 } longString) { Console.WriteLine($"Long string found: {longString}"); }
The as operator attempts to cast an object to a specified reference type, returning null if the cast fails rather than throwing an exception.
null
object value = "Hello, World!"; // Try to cast to string string message = value as string; // Check if cast was successful if (message != null) { Console.WriteLine($"Successful cast: {message}"); }
Understanding these operators helps you write more elegant, safe code when working with polymorphic types in C#.
Reading a file line by line is useful when handling large files without loading everything into memory at once.
✅ Best Practice: Use File.ReadLines() which is more memory efficient.
Example
foreach (string line in File.ReadLines("file.txt")) { Console.WriteLine(line); }
Why use ReadLines()?
Reads one line at a time, reducing overall memory usage. Ideal for large files (e.g., logs, CSVs).
Alternative: Use StreamReader (More Control)
For scenarios where you need custom processing while reading the contents of the file:
using (StreamReader reader = new StreamReader("file.txt")) { string? line; while ((line = reader.ReadLine()) != null) { Console.WriteLine(line); } }
Why use StreamReader?
Lets you handle exceptions, encoding, and buffering. Supports custom processing (e.g., search for a keyword while reading).
When to Use ReadAllLines()? If you need all lines at once, use:
string[] lines = File.ReadAllLines("file.txt");
Caution: Loads the entire file into memory—avoid for large files!
Using SqlDataReader asynchronously prevents blocking the main thread, improving performance in web apps and large queries. Here’s how to do it properly.
Use await with ExecuteReaderAsync()
using (SqlConnection conn = new SqlConnection(connectionString)) { await conn.OpenAsync(); using (SqlCommand cmd = new SqlCommand("SELECT * FROM Users", conn)) using (SqlDataReader reader = await cmd.ExecuteReaderAsync()) { while (await reader.ReadAsync()) { Console.WriteLine(reader["Username"]); } } // ✅ Auto-closes reader } // ✅ Auto-closes connection
Why use async?
A couple of reasons:
⚡ Alternative: ConfigureAwait(false) for ASP.NET
Use ConfigureAwait(false) in library code to avoid deadlocks in UI frameworks like ASP.NET.
using (SqlConnection conn = new SqlConnection(connectionString)) { await conn.OpenAsync().ConfigureAwait(false); using (SqlCommand cmd = new SqlCommand("SELECT * FROM Users", conn)) using (SqlDataReader reader = await cmd.ExecuteReaderAsync().ConfigureAwait(false)) { while (await reader.ReadAsync().ConfigureAwait(false)) { Console.WriteLine(reader["Username"]); } } }
When it comes to iterating over collections in C#, the performance difference between foreach and for loops primarily depends on the collection type being traversed.
For arrays and Lists, a traditional for loop with indexing can be marginally faster because it avoids the overhead of creating an enumerator object, especially in performance-critical scenarios.
The foreach loop internally creates an IEnumerator, which adds a small memory allocation and method call overhead.
However, for most modern applications, this performance difference is negligible and often optimized away by the JIT compiler.
The readability benefits of foreach typically outweigh the minor performance gains of for loops in non-critical code paths.
Collections like LinkedList or those implementing only IEnumerable actually perform better with foreach since they don't support efficient random access.
The rule of thumb: use foreach for readability in most cases, and only switch to for loops when benchmarking shows a meaningful performance improvement in your specific high-performance scenarios.
// Collection to iterate List<int> numbers = Enumerable.Range(1, 10000).ToList(); // Using for loop public void ForLoopExample(List<int> items) { int sum = 0; for (int i = 0; i < items.Count; i++) { sum += items[i]; } // For loop can be slightly faster for List<T> and arrays // because it avoids creating an enumerator } // Using foreach loop public void ForEachLoopExample(List<int> items) { int sum = 0; foreach (int item in items) { sum += item; } // More readable and works well for any collection type // Preferred for most scenarios where performance isn't critical } // For a LinkedList, foreach is typically faster public void LinkedListExample(LinkedList<int> linkedItems) { int sum = 0; // This would be inefficient with a for loop since LinkedList // doesn't support efficient indexing foreach (int item in linkedItems) { sum += item; } }
Register for my free weekly newsletter.