T-SQL Challenge with SQL Server 2012’s Sequence ObjectCommented on: 4 years ago
(December 16, 2010)
Most efficient way I can think of is to have the sequence INCREMENT BY 2 each time, and insert (NEXT VALUE FOR dbo.Seq1) and (NEXT VALUE FOR dbo.Seq1 + 1) into key1 and key2 (respectively) in T1....
SQL Server 2012 T-SQL at a Glance – SequencesCommented on: 5 years ago
(November 23, 2010)
I believe I'm right in saying that other than column order, the order of items in the SELECT list doesn't mean anything - so it would be wrong for SQL Server to arbitrarily provide e.g. 1 as the...
Geekiest Sig ChallengeCommented on: 5 years ago
(February 1, 2010)
Now that's just geeky to the max! If you had written a .Net app to pull out the pixel positions I'm not sure whether it would've been more or less geeky, mind. :)
Solutions to Logic Puzzle - Crossing a DesertCommented on: 5 years ago
(December 9, 2009)
The only trouble is, Itzik, that when you return to the first spot where you deposited the food you find it has been eaten by camels! Thankfully, you capture one of them, ride it for three days...
Solutions to T-SQL Challenge - Efficient Partitioned TOPCommented on: 6 years ago
(October 30, 2009)
You can cheat by relying on the randomness of the data inserted into the table an put "WHERE col1 > 95" at the end of the inner query - limiting the size of the sort. Not recommended for production...
Virtual Auxiliary Table of NumbersCommented on: 6 years ago
(October 30, 2009)
I've tried the recursive approach you suggest, oddsock and as per article #94376, having the OPTION within the function doesn't work (syntax error) - it'd be a huge pain to have to remember to...
Optimizing a Suboptimal Query PlanCommented on: 6 years ago
(March 26, 2009)
Further to my earlier reply - here are some timings (T1 w/ 1,000,000 rows, col2 'pseudorandomizedish' - SQL Server w/ 768MB RAM, testdb and tempdb on separate disks).
Original query: 29s...
Digital Magazine Archives
Browse back issues of SQL Server Pro, from January 2007 through the last issue published in April 2014. Find the back issues here.
Melissa Data's Data Quality Analyst Joseph Vertido explains how Gartner defines the critical data quality steps to prevent bad data from entering your systems in the first place, and then keep it clean over time....More
I’ll sound a bit like Captain Obvious for bringing this up, but it’s important to remember that security encompasses a lot more than protecting sensitive data from the specter of outsider threats like hackers. Properly implemented security policies also account for threat-models that include insiders – or people within your organization....More
There’s simply no way to overstate the importance of regular Transaction Log Backups. Not only do they help protect from disaster, but regular execution of T-Log backups on Full (and Bulk-Logged) Recovery databases helps keep thing “fit and trim”. Most of the time, setting up a Notification for when T-Log Backup Jobs fail is enough to let you know when something goes wrong....More