Skip to content

[Python] Non-nullable schema fields not checked in conversions from pandas #18104

@asfimport

Description

@asfimport

If you provide a schema with nullable=False but pass a DataFrame which in fact has nulls it appears the schema is ignored? I would expect an error here.

import pyarrow as pa
import pandas as pd

df = pd.DataFrame({"a":[1.2, 2.1, pd.np.NaN]})
schema = pa.schema([pa.field("a", pa.float64(), nullable=False)])
table = pa.Table.from_pandas(df, schema=schema)
table[0]

<pyarrow.lib.Column object at 0x7f213bf2fb70>
chunk 0: <pyarrow.lib.DoubleArray object at 0x7f213bf20ea8>
[
  1.2,
  2.1,
  NA
]

Reporter: Matthew Gilbert
Assignee: Wes McKinney / @wesm

PRs and other links:

Note: This issue was originally created as ARROW-2136. Please see the migration documentation for further details.

Metadata

Metadata

Assignees

Type

No type

Projects

No projects

Milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions