Bill Fellows runs into an interesting oddity:
Yet another thing that has bitten me working in SparkSQL in Databricks—this time it’s data types.
In SQL Server, a tinyint ranges from 0 to 255 but both of them allow for 256 total values. If you attempt to cast a value that doesn’t fit in that range, you’re going to raise an error.
SQL Server’s TINYINT
data type is an unsigned one-byte number, whereas TINYINT
in Spark SQL is a signed one-byte number. But that’s not the biggest difference Bill finds, so check out the post to learn more.