I am writing code in hiveQL using databricks community edition.
I have loaded the csv dataset files into dbfs, and created hive external table and tried loading table with data from uploaded dbfs file.
CREATE EXTERNAL TABLE IF NOT EXISTS data_table(
ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
result was OK.
But when I queried
**select from data_table*
It throws an exception:
Error in SQL statement: IOException: Path: /FileStore/tables/clinicaltrial_2021/mesh.csv is a directory, which is not supported by the record reader when
mapreduce.input.fileinputformat.input.dir.recursive is false.
When I checked dbfs directory:
There exists a new directory mesh.csv created by dbfs with many subdirectories with same mesh.csv
Any help is highly appreciated.