Based on this old thread :
I am thinking, probably the issue is not with the pipeline parameters or variables. Wondering , if you have Rule-Based Mapping in any of your DataFlow activity. In the rule-based mapping, you can define columns "name", "type", "stream", "origin", "position" etc. Would you please check the expression (if you have any), there is possibility that it could caused this error.
Also, to check for corruption in a .parquet file, you can use tools like Apache Parquet CLI or read the file using a Python script with libraries like pyarrow
or pandas
:
import pyarrow.parquet as pq
def check_parquet(file_path):
try:
table = pq.read_table(file_path)
print(table)
return True
except Exception as e:
print(f"Error reading parquet file: {e}")
return False
# Example usage
file_path = 'path/to/your/file.parquet'
is_valid = check_parquet(file_path)
print(f"File valid: {is_valid}")