What is the logic behind requiring the same length when using UNPIVOT?
This question may only be truly answerable by the people who worked on the implementation of UNPIVOT
. You might be able to obtain this by contacting them for support. The following is my understanding of the reasoning, which may not be 100% accurate:
T-SQL contains any number of instances of weird semantics and other counter-intuitive behaviours. Some of these will eventually go away as part of deprecation cycles, but others may never be 'improved' or 'fixed'. Quite aside from anything else, applications exist that depend on these behaviours, so backward compatibility has to be preserved.
The rules for implicit conversions, and expression type derivation account for a significant proportion of the weirdness mentioned above. I do not envy the testers who have to ensure that the weird (and often undocumented) behaviours are preserved (under all combinations of SET
session values and so on) for new versions.
That said, there is no good reason not to make improvements, and avoid past mistakes, when introducing new language features (with obviously no backward compatibility baggage). New features like recursive common table expressions (as mentioned by Andriy M in a comment) and UNPIVOT
were free to have relatively sane semantics and clearly-defined rules.
There will be a range of views as to whether including the length in the type is taking explicit typing too far, but personally I welcome it. In my view, the types varchar(25)
and varchar(50)
are not the same, any more than decimal(8)
and decimal(10)
are. Special casing string type conversion complicates things unnecessarily and adds no real value, in my opinion.
One could argue that only implicit conversions that might lose data should be required to be explicitly stated, but there are edge-cases there too. Ultimately, a conversion is going to be needed, so we might as well make it explicit.
If the implicit conversion from varchar(25)
to varchar(50)
were allowed, it would just be another (most likely hidden) implicit conversion, with all the usual weird edge cases and SET
setting sensitivities. Why not make the implementation the simplest and most explicit possible? (Nothing is perfect, however, and it is a shame that hiding varchar(25)
and varchar(50)
inside a sql_variant
is allowed.)
Rewriting the UNPIVOT
with APPLY
and UNION ALL
avoids the (better) type behaviour because the rules for UNION
are subject to backward compatibility, and are documented in Books Online as allowing different types so long as they are comparable using implicit conversion (for which the arcane rules of data type precedence are used, and so on).
The workaround involves being explicit about the data types and adding explicit conversions where necessary. This looks like progress to me :)
One way to write the explicitly-typed workaround:
SELECT
U.PersonId,
U.ColumnName,
U.Value
FROM dbo.People AS P
CROSS APPLY
(
VALUES (CONVERT(varchar(50), Lastname))
) AS CA (Lastname)
UNPIVOT
(
Value FOR
ColumnName IN (P.Firstname, CA.Lastname)
) AS U;
Recursive CTE example:
-- Fails
WITH R AS
(
SELECT Dummy = 'A row'
UNION ALL
SELECT 'Another row'
FROM R
WHERE Dummy = 'A row'
)
SELECT Dummy
FROM R;
-- Succeeds
WITH R AS
(
SELECT Dummy = CONVERT(varchar(11), 'A row')
UNION ALL
SELECT CONVERT(varchar(11), 'Another row')
FROM R
WHERE Dummy = 'A row'
)
SELECT Dummy
FROM R;
Finally, note that the rewrite using CROSS APPLY
in the question is not quite the same as the UNPIVOT
, because it does not reject NULL
attributes.
This is a session setting; there is nothing to roll back. If you had issued SET DATEFORMAT DMY;
or SET LANGUAGE FRENCH;
, would you expect an error to revert you to MDY
or english_us
? What if you had SET SHOWPLAN
on, would you expect a rollback to stop showing you plans on subsequent queries?
To be fair, the IDENTITY_INSERT
case is a little special, since it's the only session setting I can think of off the top of my head that interacts directly with a table. But it's still applicable only to your session. The thinking is probably that you could go on to insert many rows in multiple batches, even after an error has been raised (as long as it hasn't severed the connection).
As an aside, you're using SQL Server 2012, so why are you using SQL Server 7.0 error handling techniques? You could use TRY/CATCH
instead of SET XACT_ABORT
and then turn the setting off in the CATCH
if it's on (which you can check). For a good primer on error handling, see Erland Sommarskog's articles on Error and Transaction Handling in SQL Server:
Best Answer
I don't think there's any real "internals" reason. The metadata is stored at column level not table level. It would need a rethink though of scalar functions such as
scope_identity()
and pseudo column syntax such as$identity
as there would now be ambiguities.Philosophically if the purpose of
identity
is to produce something that uniquely identifies an entity why would you need two different arbitrary calculated values acting in that role?And where is the benefit anyway? This is a cross site dupe so I'll repeat my example from SO.
Since SQL Server 2012 you can knock yourself out and add as many columns using sequence defaults as you want to a table though. For example: