Additional Articles


Technical Article

SQL Server 2005: Integrating SQL, XML, and XQuery

The evolution of SQL and the XML Query Language (XQuery) continues with the work of the World Wide Web Consortium (W3C) and the InterNational Committee for Information Technology Standards (INCITS). Providers of SQL database management systems have upgraded products such as Microsoft SQL Server to support the storage and retrieval of XML documents. Microsoft has provided stored procedures and Transact-SQL extensions for working with XML. On the horizon are even more changes as Microsoft introduces SQL Server 2005. (MP3 Audio)

2005-01-13

1,665 reads

Technical Article

Understanding "Yukon" Schema Separation

Well it has finally arrived, at least in the Beta version. Microsoft's long awaited latest version of it's SQL Server product has arrived in Beta version and holds promise to be a major and successful revision of this fine product. I have had the Beta version for a few months now and one of the new security items that has intrigued me the most is the separation of users and schemas. I've worked with this form of separation before in Microsoft's chief competitor, but this article is not a comparison of the two products or the way they implement schema separation; it is an article on the basics of user/schema separation for those SQL Server DBAs who may have not worked with separated schema separation before.

2005-01-12

2,354 reads

Technical Article

Trace-scrubbing Tools

Andrew Zanevsky shares his trace-scrubbing procedures that make it easy for you to handle large trace files and aggregate transactions by type–even when captured T-SQL code has variations.

SQL Server Profiler is a veritable treasure trove when it comes to helping DBAs optimize their T-SQL code. But, the surfeit of riches (I'm reminded of the Arabian Nights tale of Aladdin) can be overwhelming. I recently had one of those "sinking" feelings when I first tried to make sense of the enormous amount of data collected by traces on a client's servers. At this particular client, the online transactions processing system executes more than 4 million database transactions per hour. That means that even a 30-minute trace that captures "SQL Batch Completed" events results in a table with 2 million rows. Of course, it's simply impractical to process so many records without some automation, and even selecting the longest or most expensive transactions doesn't necessarily help in identifying bottlenecks. After all, short transactions can be the culprits of poor performance when executed thousands of times per minute.

2005-01-11

2,003 reads

Technical Article

Manipulating Microsoft SQL Server Using SQL Injection

Focuses on advanced techniques that can be used in an attack on an application utilizing Microsoft SQL Server as a backend. These techniques demonstrate how an attacker could use a SQL Injection vulnerability to retrieve the database content from behind a firewall and penetrate the internal network. Also provided are recommendations on how to prevent such attacks.

2005-01-05

2,626 reads

Blogs

FabCon 2025 Recap

By

Hey fellow data friends! Wow, March was an absolutely crazy month (in the best...

Privacy Policy

By

We value your privacy. This policy explains what personal data we collect and why. What...

Why use XACT_ABORT?

By

This came up one day at my work when a developer was using it....

Read the latest Blogs

Forums

MasterDataservices website how to change the web application path

By jaango123

So please check the screenshot. The web application path, I just want /MDS. But...

Break string into SQL column regarding a specific character

By kakovatos7

Hi all, I want to separate string within a table column, accordint to a...

Importing Flat Files

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Importing Flat Files

Visit the forum

Question of the Day

Importing Flat Files

What types of sources can be used with the Import Flat File Wizard in SQL Server Management Studio (SSMS)?

See possible answers