June 30, 2009 at 4:31 pm
Thank you, Simon. Yes, you gave me this formula before and it works perfect.
Best regards, Irina.
November 3, 2009 at 2:09 am
Hi
I have multi value parameter:
a
b
c,e
f,m
d
The selected values are returned as b,c,e,f,m
The UDF returns 5 different items when it should return 3 items. The problem is that the values have comma in them. Is there a way to deal with this? I can not use the ID numbers.
Thanks
November 3, 2009 at 5:40 am
Naeem_Khan (11/3/2009)
HiI have multi value parameter:
a
b
c,e
f,m
d
The selected values are returned as b,c,e,f,m
The UDF returns 5 different items when it should return 3 items. The problem is that the values have comma in them. Is there a way to deal with this? I can not use the ID numbers.
Thanks
Are there, in fact, end of line characters in the multiline parameter? If so, include them in the passed data and split on them.
--Jeff Moden
Change is inevitable... Change for the better is not.
November 3, 2009 at 6:02 am
I don't think there are end of line characters.
November 12, 2009 at 8:49 pm
Naeem_Khan (11/3/2009)
I don't think there are end of line characters.
If they print on different lines, I can just about guarantee there are. π
--Jeff Moden
Change is inevitable... Change for the better is not.
August 13, 2010 at 1:44 pm
I have used this code many times since I first read this article. Now, how do I create multiple variables within the same stored proc?
August 13, 2010 at 1:49 pm
Just declare as many variables as you need, convert report parameters to delimited string send to sproc ... use the same scenario. Let me know if you have any problem.
August 13, 2010 at 1:52 pm
Thanks for the quick reply. Having a little trouble getting the order correct with the code. Get the declare and set statements to work but the CTE is tripping me up.
August 13, 2010 at 2:17 pm
INSTALL function (code below) and use it to split a value
CREATE FUNCTION [dbo].[StringToTable]
(@RawString nvarchar(4000), @Delimiter nvarchar(10) )
RETURNS
@VList TABLE (RowID int IDENTITY(1,1),SplitedValue varchar(200))
AS
BEGIN
WITH CSVCte (StartPos, EndPos) AS
(SELECT 1 AS StartPos, CHARINDEX(@Delimiter , @RawString + @Delimiter) AS EndPos
UNION ALL
SELECT EndPos + 1 AS StartPos , CHARINDEX(@Delimiter,@RawString + @Delimiter , EndPos + 1) AS EndPos
FROM CSVCTE WHERE CHARINDEX(@Delimiter, @RawString + @Delimiter, EndPos + 1) <> 0)
INSERT INTO @VList
SELECT SUBSTRING(@RawString, StartPos,EndPos - StartPos) FROM CSVCte
RETURN
END
GO
-- sample call: select SplitedValue from dbo.StringToTable('A,B,C,D,E,F,G', ',')
August 14, 2010 at 11:23 am
Alex Grinberg-230981 (8/13/2010)
INSTALL function (code below) and use it to split a valueCREATE FUNCTION [dbo].[StringToTable]
(@RawString nvarchar(4000), @Delimiter nvarchar(10) )
RETURNS
@VList TABLE (RowID int IDENTITY(1,1),SplitedValue varchar(200))
AS
BEGIN
WITH CSVCte (StartPos, EndPos) AS
(SELECT 1 AS StartPos, CHARINDEX(@Delimiter , @RawString + @Delimiter) AS EndPos
UNION ALL
SELECT EndPos + 1 AS StartPos , CHARINDEX(@Delimiter,@RawString + @Delimiter , EndPos + 1) AS EndPos
FROM CSVCTE WHERE CHARINDEX(@Delimiter, @RawString + @Delimiter, EndPos + 1) <> 0)
INSERT INTO @VList
SELECT SUBSTRING(@RawString, StartPos,EndPos - StartPos) FROM CSVCte
RETURN
END
GO
-- sample call: select SplitedValue from dbo.StringToTable('A,B,C,D,E,F,G', ',')
That's a different take on using a recursive CTE to do it. I'll have to check the performance, though, because recursive CTE's tend to be nothing more than "Hidden RBAR" with all the performance problems associated with RBAR.
--Jeff Moden
Change is inevitable... Change for the better is not.
August 14, 2010 at 12:00 pm
Ok... here we go with a test for Alex's function. I was right about the CTE still being "Hidden RBAR" even though it uses all variables and is an InLine Table Valued Function (iTVF).
Here's the data generation code again...
--=================================================================================================
-- Create a CSV test table with 10000 rows and 10 random CSV elements per row.
-- The randomness of the elements also prevents the delimiters for showing up in the same
-- position for each row. SQL Server would figure that out and cache the information making
-- some splitting techniques seem faster than they really are.
-- This section of the code takes just a couple of seconds to run because XML concatenation
-- is very fast (especially when compared to XML splitting or shredding).
-- Jeff Moden
--=================================================================================================
--===== Conditionally drop the test table to make reruns easier
IF OBJECT_ID('TempDB..#CsvTest','U') IS NOT NULL
DROP TABLE #CsvTest
--===== This creates and populates a test table on the fly containing a
-- sequential column and a randomly generated CSV Parameter column.
SELECT TOP (10000) --Controls the number of rows in the test table
ISNULL(ROW_NUMBER() OVER (ORDER BY(SELECT NULL)),0) AS RowNum,
(
SELECT CAST(STUFF( --=== STUFF get's rid of the leading comma
( --=== This builds CSV row with a leading comma
SELECT TOP (10) --Controls the number of CSV elements in each row
','+CAST(ABS(CHECKSUM(NEWID()))%100000 AS VARCHAR(10))
FROM Master.sys.All_Columns ac3 --Classic cross join pseudo-cursor
CROSS JOIN Master.sys.All_Columns ac4 --can produce row sets up 16 million
WHERE ac3.Object_ID = ac1.Object_ID --Without this, all rows would be the same
FOR XML PATH('')
)
,1,1,'') AS VARCHAR(8000))
) AS CsvParameter
INTO #CsvTest
FROM Master.sys.All_Columns ac1 --Classic cross join pseudo-cursor
CROSS JOIN Master.sys.All_Columns ac2 --can produce row sets up 16 million
--===== Let's add a PK just for grins. Since it's a temp table, we won't name it.
ALTER TABLE #CsvTest
ADD PRIMARY KEY CLUSTERED (RowNum) WITH FILLFACTOR = 100
--===== Sanity check... let's see what we've got in the test table
SELECT * FROM #CsvTest
Here's the test code comparing the method I've demo'd several times with Alex's recursive CTE...
--===== DelimitedSplit8k solution
SELECT test.RowNum, split.ItemNumber, split.Item
INTO #Result1
FROM #CsvTest test
CROSS APPLY
(
SELECT ItemNumber, Item
FROM dbo.DelimitedSplit8k(test.CsvParameter,',')
) split
;
GO
--===== Alex's Recursive CTE solution
SELECT test.RowNum, split.RowID, split.SplitedValue
INTO #Result2
FROM #CsvTest test
CROSS APPLY
(
SELECT RowID, SplitedValue
FROM [StringToTable](test.CsvParameter,',')
) split
;
GO
And, here are the results from the Profiler run... I'd strongly recommend that recursive CTE methods be avoided at all costs because they cost so much and are so very slow... this difference was only on 10,000 rows by 10 elements (100,000 rows total output) which can easily represent the number of hits on a busy server in a second or two.
Of course, this is one of the places where a CLR comes in handy. It'll be about twice as fast in this case.
--Jeff Moden
Change is inevitable... Change for the better is not.
August 14, 2010 at 12:09 pm
Ah... sorry, my apologies. I thought I'd posted the 8k splitter code on this thread before. Guess not. So, here it is. It's mostly comments to explain the history and operation of the code. The code itself is actually very short and very, very fast...
CREATE FUNCTION dbo.DelimitedSplit8K
/***************************************************************************************************
Purpose:
Split a given string at a given delimiter and return a list of the split elements (items).
Returns:
iTVF containing the following:
ItemNumber = Element position of Item as a BIGINT (not converted to INT to eliminate a CAST)
Item = Element value as a VARCHAR(8000)
CROSS APPLY Usage Example:
---------------------------------------------------------------------------------------------------
--===== Conditionally drop the test tables to make reruns easier for testing.
-- (this is NOT a part of the solution)
IF OBJECT_ID('tempdb..#JBMTest') IS NOT NULL
DROP TABLE #JBMTest
;
--===== Create and populate a test table on the fly (this is NOT a part of the solution).
SELECT *
INTO #JBMTest
FROM (
SELECT 1,'1,10,100,1000,10000,100000,1000000' UNION ALL
SELECT 2,'2000000,200000,20000,2000,200,20,2' UNION ALL
SELECT 3, 'This,is,a,test' UNION ALL
SELECT 4, 'and so is this' UNION ALL
SELECT 5, 'This, too (no pun intended)'
) d (SomeID,SomeValue)
;
GO
--===== Split the CSV column for the whole table using CROSS APPLY (this is the solution)
SELECT test.SomeID, split.ItemNumber, split.Item
FROM #JBMTest test
CROSS APPLY
(
SELECT ItemNumber, Item
FROM dbo.DelimitedSplit8k(test.SomeValue,',')
) split
;
---------------------------------------------------------------------------------------------------
Notes:
1. Optimized for VARCHAR(7999) or less. No testing or error reporting for truncation at 7999
characters is done.
2. Optimized for single character delimiter. Multi-character delimiters should be resolved
externally from this function.
3. Optimized for use with CROSS APPLY.
4. Does not "trim" elements just in case leading or trailing blanks are intended.
5. If you don't know how a Tally table can be used to replace loops, please see the following...
http://www.sqlservercentral.com/articles/T-SQL/62867/
6. Changing this function to use VARCHAR(MAX) will cause it to run twice as slow. It's just the
nature of VARCHAR(MAX) whether it fits in-row or not.
7. Multi-machine testing for the method of using UNPIVOT instead of 10 SELECT/UNION ALLs shows
that the UNPIVOT method is quite machine dependent and can slow things down quite a bit.
8. Performance testing shows using "TOP" for the limiting criteria of "N" is actually
slower and slightly more CPU intensive than the traditional WHERE N < LEN(@pString) + 2.
9. Performance testing shows using ORDER BY (SELECT x) where "x" is anything is actually
slower and slightly more CPU intensive than the traditional ORDER BY (SELECT N).
Credits:
This code is the product of many people's efforts including but not limited to the following:
cteTally concept originally by Iztek Ben Gan and "decimalized" by Lynn Pettis (and others) for a
bit of extra speed and finally redacted by Jeff Moden for a different slant on readability and
compactness. Hat's off to Paul White for his simple explanations of CROSS APPLY. Finally,
special thanks to Erland Sommarskog for his tireless efforts to help people understand
what you can actually do with T-SQL. I also thank whoever wrote the first article I ever saw
on "numbers tables" which is located at the following URL ...
Revision History:
Rev 00 - 20 Jan 2010 - Concept: Lynn Pettis and others.
Redaction/Implementation: Jeff Moden
- Base 10 redaction and reduction for CTE. (Total rewrite)
Rev 01 - 13 Mar 2010 - Jeff Moden
- Removed one additional concatenation and one subtraction from the SUBSTRING in the
SELECT List for that tiny bit of extra speed.
Rev 02 - 14 Apr 2010 - Jeff Moden
- No code changes. Added CROSS APPLY usage example to the header, some additional credits,
and extra documentation.
Rev 03 - 18 Apr 2010 - Jeff Moden
- No code changes. Added notes 7, 8, and 9 about certain "optimizations" that dont'
actually work for this type of function.
Rev 04 - 29 Jun 2010 - Jeff Moden
- Added WITH SCHEMABINDING thanks to a note by Paul White. This prevents an unnecessary
"Table Spool" when the function is used in an UPDATE statement even though the function
makes no external references.
***************************************************************************************************/
--===== Define I/O parameters
(
@pString VARCHAR(7999),
@pDelimiter CHAR(1)
)
RETURNS TABLE
WITH SCHEMABINDING
AS
RETURN
--===== "Inline" CTE Driven "Tally Tableβ produces values up to
-- 10,000... enough to cover VARCHAR(8000)
WITH
E1(N) AS ( --=== Create Ten 1's
SELECT 1 UNION ALL SELECT 1 UNION ALL
SELECT 1 UNION ALL SELECT 1 UNION ALL
SELECT 1 UNION ALL SELECT 1 UNION ALL
SELECT 1 UNION ALL SELECT 1 UNION ALL
SELECT 1 UNION ALL SELECT 1 --10
),
E2(N) AS (SELECT 1 FROM E1 a, E1 b), --100
E4(N) AS (SELECT 1 FROM E2 a, E2 b), --10,000
cteTally(N) AS (SELECT ROW_NUMBER() OVER (ORDER BY (SELECT N)) FROM E4)
--===== Do the split
SELECT ROW_NUMBER() OVER (ORDER BY N) AS ItemNumber,
SUBSTRING(@pString, N, CHARINDEX(@pDelimiter, @pString + @pDelimiter, N) - N) AS Item
FROM cteTally
WHERE N <= LEN(@pString)
AND SUBSTRING(@pDelimiter + @pString, N, 1) = @pDelimiter
;
GO
--Jeff Moden
Change is inevitable... Change for the better is not.
August 16, 2010 at 9:53 am
Thanks Jeff for your help! Code working pretty well. Only issue I'm having is getting one parameter to be dependent on the other in a stored proc that populates a drop down list in Reporting Services.
August 16, 2010 at 8:20 pm
gary.bland (8/16/2010)
Thanks Jeff for your help! Code working pretty well. Only issue I'm having is getting one parameter to be dependent on the other in a stored proc that populates a drop down list in Reporting Services.
Thanks for the feedback, Gary. I just don't know what you mean by "getting one parameter to be dependent on the other" in this case without an example. And, remember, I admitted to not even knowing how to spell "SSRS" so I'll need a pretty good example of what you want the stored proc to do. π
--Jeff Moden
Change is inevitable... Change for the better is not.
August 17, 2010 at 11:09 am
What I am trying to accomplish is when I choose a parameter I need to pass the result into another parameter in order to limit the number of available selections. An example would be State and City. Selecting a State would then limit the selection of Cities.
Viewing 15 posts - 46 through 60 (of 69 total)
You must be logged in to reply to this topic. Login to reply