How to speed up a PostgreSQL group by query through multiple joins? - postgresql
This query searches for product_groupings often purchased with product_grouping ID 99999. As this query fans out to all the orders that contain product_grouping 99999, and then joins back down to count the number of times each product_grouping has been ordered, and takes the top 10.
Is there any way to speed this query up?
SELECT product_groupings.*, count(product_groupings.id) AS product_groupings_count
FROM "product_groupings"
INNER JOIN "products" ON "product_groupings"."id" = "products"."product_grouping_id"
INNER JOIN "variants" ON "products"."id" = "variants"."product_id"
INNER JOIN "order_items" ON "variants"."id" = "order_items"."variant_id"
INNER JOIN "shipments" ON "order_items"."shipment_id" = "shipments"."id"
INNER JOIN "orders" ON "shipments"."order_id" = "orders"."id"
INNER JOIN "shipments" "shipments_often_purchased_with_join" ON "orders"."id" = "shipments_often_purchased_with_join"."order_id"
INNER JOIN "order_items" "order_items_often_purchased_with_join" ON "shipments_often_purchased_with_join"."id" = "order_items_often_purchased_with_join"."shipment_id"
INNER JOIN "variants" "variants_often_purchased_with_join" ON "order_items_often_purchased_with_join"."variant_id" = "variants_often_purchased_with_join"."id"
INNER JOIN "products" "products_often_purchased_with_join" ON "variants_often_purchased_with_join"."product_id" = "products_often_purchased_with_join"."id"
WHERE "products_often_purchased_with_join"."product_grouping_id" = 99999 AND (product_groupings.id != 99999) AND "product_groupings"."state" = 'active' AND ("shipments"."state" NOT IN ('pending', 'cancelled'))
GROUP BY product_groupings.id
ORDER BY product_groupings_count desc LIMIT 10
schema:
CREATE TABLE product_groupings (
id integer NOT NULL,
state character varying(255) DEFAULT 'active'::character varying,
brand_id integer,
product_content_id integer,
hierarchy_category_id integer,
hierarchy_subtype_id integer,
hierarchy_type_id integer,
product_type_id integer,
description text,
keywords text,
created_at timestamp without time zone,
updated_at timestamp without time zone
);
CREATE INDEX index_product_groupings_on_brand_id ON product_groupings USING btree (brand_id);
CREATE INDEX index_product_groupings_on_hierarchy_category_id ON product_groupings USING btree (hierarchy_category_id);
CREATE INDEX index_product_groupings_on_hierarchy_subtype_id ON product_groupings USING btree (hierarchy_subtype_id);
CREATE INDEX index_product_groupings_on_hierarchy_type_id ON product_groupings USING btree (hierarchy_type_id);
CREATE INDEX index_product_groupings_on_name ON product_groupings USING btree (name);
CREATE INDEX index_product_groupings_on_product_content_id ON product_groupings USING btree (product_content_id);
CREATE INDEX index_product_groupings_on_product_type_id ON product_groupings USING btree (product_type_id);
ALTER TABLE ONLY product_groupings
ADD CONSTRAINT product_groupings_pkey PRIMARY KEY (id);
CREATE TABLE products (
id integer NOT NULL,
name character varying(255) NOT NULL,
prototype_id integer,
deleted_at timestamp without time zone,
created_at timestamp without time zone,
updated_at timestamp without time zone,
item_volume character varying(255),
upc character varying(255),
state character varying(255),
volume_unit character varying(255),
volume_value numeric,
container_type character varying(255),
container_count integer,
upc_ext character varying(8),
product_grouping_id integer,
short_pack_size character varying(255),
short_volume character varying(255),
additional_upcs character varying(255)[] DEFAULT '{}'::character varying[]
);
CREATE INDEX index_products_on_additional_upcs ON products USING gin (additional_upcs);
CREATE INDEX index_products_on_deleted_at ON products USING btree (deleted_at);
CREATE INDEX index_products_on_name ON products USING btree (name);
CREATE INDEX index_products_on_product_grouping_id ON products USING btree (product_grouping_id);
CREATE INDEX index_products_on_prototype_id ON products USING btree (prototype_id);
CREATE INDEX index_products_on_upc ON products USING btree (upc);
ALTER TABLE ONLY products
ADD CONSTRAINT products_pkey PRIMARY KEY (id);
CREATE TABLE variants (
id integer NOT NULL,
product_id integer NOT NULL,
sku character varying(255) NOT NULL,
name character varying(255),
price numeric(8,2) DEFAULT 0.0 NOT NULL,
deleted_at timestamp without time zone,
supplier_id integer,
created_at timestamp without time zone,
updated_at timestamp without time zone,
inventory_id integer,
product_active boolean DEFAULT false NOT NULL,
original_name character varying(255),
original_item_volume character varying(255),
protected boolean DEFAULT false NOT NULL,
sale_price numeric(8,2) DEFAULT 0.0 NOT NULL
);
CREATE INDEX index_variants_on_inventory_id ON variants USING btree (inventory_id);
CREATE INDEX index_variants_on_product_id_and_deleted_at ON variants USING btree (product_id, deleted_at);
CREATE INDEX index_variants_on_sku ON variants USING btree (sku);
CREATE INDEX index_variants_on_state_attributes ON variants USING btree (deleted_at, product_active, protected, id);
CREATE INDEX index_variants_on_supplier_id ON variants USING btree (supplier_id);
ALTER TABLE ONLY variants
ADD CONSTRAINT variants_pkey PRIMARY KEY (id);
CREATE TABLE order_items (
id integer NOT NULL,
price numeric(8,2),
total numeric(8,2),
variant_id integer NOT NULL,
shipment_id integer,
created_at timestamp without time zone,
updated_at timestamp without time zone,
quantity integer DEFAULT 1
);
CREATE INDEX index_order_items_on_shipment_id ON order_items USING btree (shipment_id);
CREATE INDEX index_order_items_on_variant_id ON order_items USING btree (variant_id);
ALTER TABLE ONLY order_items
ADD CONSTRAINT order_items_pkey PRIMARY KEY (id);
CREATE TABLE shipments (
id integer NOT NULL,
order_id integer,
shipping_method_id integer NOT NULL,
number character varying,
state character varying(255) DEFAULT 'pending'::character varying NOT NULL,
created_at timestamp without time zone,
updated_at timestamp without time zone,
supplier_id integer,
confirmed_at timestamp without time zone,
canceled_at timestamp without time zone,
out_of_hours boolean DEFAULT false NOT NULL,
delivered_at timestamp without time zone,
uuid uuid DEFAULT uuid_generate_v4()
);
CREATE INDEX index_shipments_on_order_id_and_supplier_id ON shipments USING btree (order_id, supplier_id);
CREATE INDEX index_shipments_on_state ON shipments USING btree (state);
CREATE INDEX index_shipments_on_supplier_id ON shipments USING btree (supplier_id);
ALTER TABLE ONLY shipments
ADD CONSTRAINT shipments_pkey PRIMARY KEY (id);
CREATE TABLE orders (
id integer NOT NULL,
number character varying(255),
ip_address character varying(255),
state character varying(255),
ship_address_id integer,
active boolean DEFAULT true NOT NULL,
completed_at timestamp without time zone,
created_at timestamp without time zone,
updated_at timestamp without time zone,
tip_amount numeric(8,2) DEFAULT 0.0,
confirmed_at timestamp without time zone,
delivery_notes text,
cancelled_at timestamp without time zone,
courier boolean DEFAULT false NOT NULL,
scheduled_for timestamp without time zone,
client character varying(255),
subscription_id character varying(255),
pickup_detail_id integer,
);
CREATE INDEX index_orders_on_bill_address_id ON orders USING btree (bill_address_id);
CREATE INDEX index_orders_on_completed_at ON orders USING btree (completed_at);
CREATE UNIQUE INDEX index_orders_on_number ON orders USING btree (number);
CREATE INDEX index_orders_on_ship_address_id ON orders USING btree (ship_address_id);
CREATE INDEX index_orders_on_state ON orders USING btree (state);
ALTER TABLE ONLY orders
ADD CONSTRAINT orders_pkey PRIMARY KEY (id);
Query plan:
Limit (cost=685117.80..685117.81 rows=10 width=595) (actual time=33659.659..33659.661 rows=10 loops=1)
Output: product_groupings.id, product_groupings.featured, product_groupings.searchable, product_groupings.state, product_groupings.brand_id, product_groupings.product_content_id, product_groupings.hierarchy_category_id, product_groupings.hierarchy_subtype_id, product_groupings.hierarchy_type_id, product_groupings.product_type_id, product_groupings.meta_description, product_groupings.meta_keywords, product_groupings.name, product_groupings.permalink, product_groupings.description, product_groupings.keywords, product_groupings.created_at, product_groupings.updated_at, product_groupings.tax_category_id, product_groupings.trimmed_name, (count(product_groupings.id))
Buffers: shared hit=259132 read=85657, temp read=30892 written=30886
I/O Timings: read=5542.213
-> Sort (cost=685117.80..685117.81 rows=14 width=595) (actual time=33659.658..33659.659 rows=10 loops=1)
Output: product_groupings.id, product_groupings.featured, product_groupings.searchable, product_groupings.state, product_groupings.brand_id, product_groupings.product_content_id, product_groupings.hierarchy_category_id, product_groupings.hierarchy_subtype_id, product_groupings.hierarchy_type_id, product_groupings.product_type_id, product_groupings.meta_description, product_groupings.meta_keywords, product_groupings.name, product_groupings.permalink, product_groupings.description, product_groupings.keywords, product_groupings.created_at, product_groupings.updated_at, product_groupings.tax_category_id, product_groupings.trimmed_name, (count(product_groupings.id))
Sort Key: (count(product_groupings.id))
Sort Method: top-N heapsort Memory: 30kB
Buffers: shared hit=259132 read=85657, temp read=30892 written=30886
I/O Timings: read=5542.213
-> HashAggregate (cost=685117.71..685117.75 rows=14 width=595) (actual time=33659.407..33659.491 rows=122 loops=1)
Output: product_groupings.id, product_groupings.featured, product_groupings.searchable, product_groupings.state, product_groupings.brand_id, product_groupings.product_content_id, product_groupings.hierarchy_category_id, product_groupings.hierarchy_subtype_id, product_groupings.hierarchy_type_id, product_groupings.product_type_id, product_groupings.meta_description, product_groupings.meta_keywords, product_groupings.name, product_groupings.permalink, product_groupings.description, product_groupings.keywords, product_groupings.created_at, product_groupings.updated_at, product_groupings.tax_category_id, product_groupings.trimmed_name, count(product_groupings.id)
Buffers: shared hit=259129 read=85657, temp read=30892 written=30886
I/O Timings: read=5542.213
-> Hash Join (cost=453037.24..685117.69 rows=14 width=595) (actual time=26019.889..33658.886 rows=181 loops=1)
Output: product_groupings.id, product_groupings.featured, product_groupings.searchable, product_groupings.state, product_groupings.brand_id, product_groupings.product_content_id, product_groupings.hierarchy_category_id, product_groupings.hierarchy_subtype_id, product_groupings.hierarchy_type_id, product_groupings.product_type_id, product_groupings.meta_description, product_groupings.meta_keywords, product_groupings.name, product_groupings.permalink, product_groupings.description, product_groupings.keywords, product_groupings.created_at, product_groupings.updated_at, product_groupings.tax_category_id, product_groupings.trimmed_name
Hash Cond: (order_items_often_purchased_with_join.variant_id = variants_often_purchased_with_join.id)
Buffers: shared hit=259129 read=85657, temp read=30892 written=30886
I/O Timings: read=5542.213
-> Hash Join (cost=452970.37..681530.70 rows=4693428 width=599) (actual time=22306.463..32908.056 rows=8417034 loops=1)
Output: product_groupings.id, product_groupings.featured, product_groupings.searchable, product_groupings.state, product_groupings.brand_id, product_groupings.product_content_id, product_groupings.hierarchy_category_id, product_groupings.hierarchy_subtype_id, product_groupings.hierarchy_type_id, product_groupings.product_type_id, product_groupings.meta_description, product_groupings.meta_keywords, product_groupings.name, product_groupings.permalink, product_groupings.description, product_groupings.keywords, product_groupings.created_at, product_groupings.updated_at, product_groupings.tax_category_id, product_groupings.trimmed_name, order_items_often_purchased_with_join.variant_id
Hash Cond: (products.product_grouping_id = product_groupings.id)
Buffers: shared hit=259080 read=85650, temp read=30892 written=30886
I/O Timings: read=5540.529
-> Hash Join (cost=381952.28..493289.49 rows=5047613 width=8) (actual time=21028.128..25416.504 rows=8417518 loops=1)
Output: products.product_grouping_id, order_items_often_purchased_with_join.variant_id
Hash Cond: (order_items_often_purchased_with_join.shipment_id = shipments_often_purchased_with_join.id)
Buffers: shared hit=249520 read=77729
I/O Timings: read=5134.878
-> Seq Scan on public.order_items order_items_often_purchased_with_join (cost=0.00..82689.54 rows=4910847 width=8) (actual time=0.003..1061.456 rows=4909856 loops=1)
Output: order_items_often_purchased_with_join.shipment_id, order_items_often_purchased_with_join.variant_id
Buffers: shared hit=67957
-> Hash (cost=373991.27..373991.27 rows=2274574 width=8) (actual time=21027.220..21027.220 rows=2117538 loops=1)
Output: products.product_grouping_id, shipments_often_purchased_with_join.id
Buckets: 262144 Batches: 1 Memory Usage: 82717kB
Buffers: shared hit=181563 read=77729
I/O Timings: read=5134.878
-> Hash Join (cost=249781.35..373991.27 rows=2274574 width=8) (actual time=10496.552..20383.404 rows=2117538 loops=1)
Output: products.product_grouping_id, shipments_often_purchased_with_join.id
Hash Cond: (shipments.order_id = orders.id)
Buffers: shared hit=181563 read=77729
I/O Timings: read=5134.878
-> Hash Join (cost=118183.04..233677.13 rows=1802577 width=8) (actual time=6080.516..14318.439 rows=1899610 loops=1)
Output: products.product_grouping_id, shipments.order_id
Hash Cond: (variants.product_id = products.id)
Buffers: shared hit=107220 read=55876
I/O Timings: read=5033.540
-> Hash Join (cost=83249.21..190181.06 rows=1802577 width=8) (actual time=4526.391..11330.434 rows=1899808 loops=1)
Output: variants.product_id, shipments.order_id
Hash Cond: (order_items.variant_id = variants.id)
Buffers: shared hit=88026 read=44439
I/O Timings: read=4009.465
-> Hash Join (cost=40902.30..138821.27 rows=1802577 width=8) (actual time=3665.477..8553.803 rows=1899816 loops=1)
Output: order_items.variant_id, shipments.order_id
Hash Cond: (order_items.shipment_id = shipments.id)
Buffers: shared hit=56654 read=43022
I/O Timings: read=3872.065
-> Seq Scan on public.order_items (cost=0.00..82689.54 rows=4910847 width=8) (actual time=0.003..2338.108 rows=4909856 loops=1)
Output: order_items.variant_id, order_items.shipment_id
Buffers: shared hit=55987 read=11970
I/O Timings: read=1059.971
-> Hash (cost=38059.31..38059.31 rows=812284 width=8) (actual time=3664.973..3664.973 rows=834713 loops=1)
Output: shipments.id, shipments.order_id
Buckets: 131072 Batches: 1 Memory Usage: 32606kB
Buffers: shared hit=667 read=31052
I/O Timings: read=2812.094
-> Seq Scan on public.shipments (cost=0.00..38059.31 rows=812284 width=8) (actual time=0.017..3393.420 rows=834713 loops=1)
Output: shipments.id, shipments.order_id
Filter: ((shipments.state)::text <> ALL ('{pending,cancelled}'::text[]))
Rows Removed by Filter: 1013053
Buffers: shared hit=667 read=31052
I/O Timings: read=2812.094
-> Hash (cost=37200.34..37200.34 rows=1470448 width=8) (actual time=859.887..859.887 rows=1555657 loops=1)
Output: variants.product_id, variants.id
Buckets: 262144 Batches: 1 Memory Usage: 60768kB
Buffers: shared hit=31372 read=1417
I/O Timings: read=137.400
-> Seq Scan on public.variants (cost=0.00..37200.34 rows=1470448 width=8) (actual time=0.009..479.528 rows=1555657 loops=1)
Output: variants.product_id, variants.id
Buffers: shared hit=31372 read=1417
I/O Timings: read=137.400
-> Hash (cost=32616.92..32616.92 rows=661973 width=8) (actual time=1553.664..1553.664 rows=688697 loops=1)
Output: products.product_grouping_id, products.id
Buckets: 131072 Batches: 1 Memory Usage: 26903kB
Buffers: shared hit=19194 read=11437
I/O Timings: read=1024.075
-> Seq Scan on public.products (cost=0.00..32616.92 rows=661973 width=8) (actual time=0.011..1375.757 rows=688697 loops=1)
Output: products.product_grouping_id, products.id
Buffers: shared hit=19194 read=11437
I/O Timings: read=1024.075
-> Hash (cost=125258.00..125258.00 rows=1811516 width=12) (actual time=4415.081..4415.081 rows=1847746 loops=1)
Output: orders.id, shipments_often_purchased_with_join.order_id, shipments_often_purchased_with_join.id
Buckets: 262144 Batches: 1 Memory Usage: 79396kB
Buffers: shared hit=74343 read=21853
I/O Timings: read=101.338
-> Hash Join (cost=78141.12..125258.00 rows=1811516 width=12) (actual time=1043.228..3875.433 rows=1847746 loops=1)
Output: orders.id, shipments_often_purchased_with_join.order_id, shipments_often_purchased_with_join.id
Hash Cond: (shipments_often_purchased_with_join.order_id = orders.id)
Buffers: shared hit=74343 read=21853
I/O Timings: read=101.338
-> Seq Scan on public.shipments shipments_often_purchased_with_join (cost=0.00..37153.55 rows=1811516 width=8) (actual time=0.006..413.785 rows=1847766 loops=1)
Output: shipments_often_purchased_with_join.order_id, shipments_often_purchased_with_join.id
Buffers: shared hit=31719
-> Hash (cost=70783.52..70783.52 rows=2102172 width=4) (actual time=1042.239..1042.239 rows=2097229 loops=1)
Output: orders.id
Buckets: 262144 Batches: 1 Memory Usage: 73731kB
Buffers: shared hit=42624 read=21853
I/O Timings: read=101.338
-> Seq Scan on public.orders (cost=0.00..70783.52 rows=2102172 width=4) (actual time=0.012..553.606 rows=2097229 loops=1)
Output: orders.id
Buffers: shared hit=42624 read=21853
I/O Timings: read=101.338
-> Hash (cost=20222.66..20222.66 rows=637552 width=595) (actual time=1278.121..1278.121 rows=626176 loops=1)
Output: product_groupings.id, product_groupings.featured, product_groupings.searchable, product_groupings.state, product_groupings.brand_id, product_groupings.product_content_id, product_groupings.hierarchy_category_id, product_groupings.hierarchy_subtype_id, product_groupings.hierarchy_type_id, product_groupings.product_type_id, product_groupings.meta_description, product_groupings.meta_keywords, product_groupings.name, product_groupings.permalink, product_groupings.description, product_groupings.keywords, product_groupings.created_at, product_groupings.updated_at, product_groupings.tax_category_id, product_groupings.trimmed_name
Buckets: 16384 Batches: 4 Memory Usage: 29780kB
Buffers: shared hit=9559 read=7921, temp written=10448
I/O Timings: read=405.651
-> Seq Scan on public.product_groupings (cost=0.00..20222.66 rows=637552 width=595) (actual time=0.020..873.844 rows=626176 loops=1)
Output: product_groupings.id, product_groupings.featured, product_groupings.searchable, product_groupings.state, product_groupings.brand_id, product_groupings.product_content_id, product_groupings.hierarchy_category_id, product_groupings.hierarchy_subtype_id, product_groupings.hierarchy_type_id, product_groupings.product_type_id, product_groupings.meta_description, product_groupings.meta_keywords, product_groupings.name, product_groupings.permalink, product_groupings.description, product_groupings.keywords, product_groupings.created_at, product_groupings.updated_at, product_groupings.tax_category_id, product_groupings.trimmed_name
Filter: ((product_groupings.id <> 99999) AND ((product_groupings.state)::text = 'active'::text))
Rows Removed by Filter: 48650
Buffers: shared hit=9559 read=7921
I/O Timings: read=405.651
-> Hash (cost=66.86..66.86 rows=4 width=4) (actual time=2.223..2.223 rows=30 loops=1)
Output: variants_often_purchased_with_join.id
Buckets: 1024 Batches: 1 Memory Usage: 2kB
Buffers: shared hit=49 read=7
I/O Timings: read=1.684
-> Nested Loop (cost=0.17..66.86 rows=4 width=4) (actual time=0.715..2.211 rows=30 loops=1)
Output: variants_often_purchased_with_join.id
Buffers: shared hit=49 read=7
I/O Timings: read=1.684
-> Index Scan using index_products_on_product_grouping_id on public.products products_often_purchased_with_join (cost=0.08..5.58 rows=2 width=4) (actual time=0.074..0.659 rows=6 loops=1)
Output: products_often_purchased_with_join.id
Index Cond: (products_often_purchased_with_join.product_grouping_id = 99999)
Buffers: shared hit=5 read=4
I/O Timings: read=0.552
-> Index Scan using index_variants_on_product_id_and_deleted_at on public.variants variants_often_purchased_with_join (cost=0.09..30.60 rows=15 width=8) (actual time=0.222..0.256 rows=5 loops=6)
Output: variants_often_purchased_with_join.id, variants_often_purchased_with_join.product_id
Index Cond: (variants_often_purchased_with_join.product_id = products_often_purchased_with_join.id)
Buffers: shared hit=44 read=3
I/O Timings: read=1.132
Total runtime: 33705.142 ms
Gained a significant ~20x increase in throughput using a sub select;
SELECT product_groupings.*, count(product_groupings.id) AS product_groupings_count
FROM "product_groupings"
INNER JOIN "products" ON "products"."product_grouping_id" = "product_groupings"."id"
INNER JOIN "variants" ON "variants"."product_id" = "products"."id"
INNER JOIN "order_items" ON "order_items"."variant_id" = "variants"."id"
INNER JOIN "shipments" ON "shipments"."id" = "order_items"."shipment_id"
WHERE ("product_groupings"."id" != 99999)
AND "product_groupings"."state" = 'active'
AND ("shipments"."state" NOT IN ('pending', 'cancelled'))
AND ("shipments"."order_id" IN (
SELECT "shipments"."order_id"
FROM "shipments"
INNER JOIN "order_items" ON "order_items"."shipment_id" = "shipments"."id"
INNER JOIN "variants" ON "variants"."id" = "order_items"."variant_id"
INNER JOIN "products" ON "products"."id" = "variants"."product_id"
WHERE "products"."product_grouping_id" = 99999 AND ("shipments"."state" NOT IN ('pending', 'cancelled'))
GROUP BY "shipments"."order_id"
ORDER BY "shipments"."order_id" ASC
))
GROUP BY product_groupings.id
ORDER BY product_groupings_count desc
LIMIT 10
Although I'd welcome any further optimisations. :)
Related
Postgres Optimiser using parallel sequential scan instead of index scan
I have two tables master and a child table. So for the question I will refer master as m and child table as c. Master table properties: Records: 50 million primary key (index): m_id btree index: modified_id Child table properties (1-n relationship): Records: 400 million primary key (index): c_id foreign key (btree index): m_id Query plan Gather (cost=9159.80..6768939.59 rows=18940 width=107) (actual time=137160.885..297009.782 rows=25 loops=1) Output: m.<date_column>, m.m_id, m.other_unique_id_1, m.modified_id, c.<date_column>, c.c_id, c.m_id, c.<other_column_1>, c.<other_column_2>, c.<other_column_3>, c.<other_column_4>, c.<other_column_5>, c.<other_column_6>, c.<other_column_7>, c.<other_column_8> Workers Planned: 2 Workers Launched: 2 Buffers: shared hit=11266 read=4680890 -> Hash Join (cost=8159.80..6766045.59 rows=7892 width=107) (actual time=87154.268..297003.331 rows=8 loops=3) Output: c.<date_column>, c.m_id, c.other_unique_id_1, c.modified_id, c.<date_column>, c.c_id, c.m_id, c.<other_column_1>, c.<other_column_2>, c.<other_column_3>, c.<other_column_4>, c.<other_column_5>, c.<other_column_6>, c.<other_column_7>, c.<other_column_8> Inner Unique: true Hash Cond: (c.m_id = c.m_id) Buffers: shared hit=11266 read=4680890 Worker 0: actual time=82162.730..297002.014 rows=13 loops=1 Buffers: shared hit=3789 read=1555017 Worker 1: actual time=42139.478..297002.514 rows=8 loops=1 Buffers: shared hit=3634 read=1569261 -> Parallel Seq Scan on child c (cost=0.00..6328357.20 rows=163629920 width=87) (actual time=0.638..279084.058 rows=130858840 loops=3) Output: c.<date_column>, c.c_id, c.m_id, c.<other_column_1>, c.<other_column_2>, c.<other_column_3>, c.<other_column_4>, c.<other_column_5>, c.<other_column_6>, c.<other_column_7>, c.<other_column_8> Buffers: shared hit=11171 read=4680887 Worker 0: actual time=0.814..279170.920 rows=130423418 loops=1 Buffers: shared hit=3746 read=1555017 Worker 1: actual time=1.095..278955.836 rows=131603000 loops=1 Buffers: shared hit=3592 read=1569260 -> Hash (cost=8132.89..8132.89 rows=2153 width=20) (actual time=0.046..0.047 rows=8 loops=3) Output: c.<date_column>, c.m_id, c.other_unique_id_1, c.modified_id Buckets: 4096 Batches: 1 Memory Usage: 33kB Buffers: shared hit=35 read=3 Worker 0: actual time=0.026..0.027 rows=8 loops=1 Buffers: shared hit=13 Worker 1: actual time=0.059..0.060 rows=8 loops=1 Buffers: shared hit=12 read=1 -> Bitmap Heap Scan on master c (cost=41.25..8132.89 rows=2153 width=20) (actual time=0.031..0.041 rows=8 loops=3) Output: c.<date_column>, c.m_id, c.other_unique_id_1, c.modified_id Recheck Cond: (c.modified_id = 561869) Heap Blocks: exact=8 Buffers: shared hit=35 read=3 Worker 0: actual time=0.018..0.024 rows=8 loops=1 Buffers: shared hit=13 Worker 1: actual time=0.043..0.055 rows=8 loops=1 Buffers: shared hit=12 read=1 -> Bitmap Index Scan on ix_master_modified_id (cost=0.00..40.71 rows=2153 width=0) (actual time=0.027..0.027 rows=8 loops=3) Index Cond: (c.modified_id = 561869) Buffers: shared hit=11 read=3 Worker 0: actual time=0.015..0.015 rows=8 loops=1 Buffers: shared hit=5 Worker 1: actual time=0.038..0.038 rows=8 loops=1 Buffers: shared hit=4 read=1 Planning time: 0.354 ms Execution time: 297009.825 ms Query select * from master m inner join child c on m.m_id = c.m_id where m.modified_id = <xyz> DDL statement for Master table CREATE TABLE master ( <date_column> timestamp NULL DEFAULT CURRENT_DATE, m_id serial4 NOT NULL, <other_unique_id_1> int4 NULL, modified_id int4 NULL, CONSTRAINT master_pkey PRIMARY KEY (m_id), CONSTRAINT <foreign key> FOREIGN KEY (<other_unique_id_1>) REFERENCES <other table>(<other_unique_id_1>) ); CREATE INDEX ix_master_modified_id ON master USING btree (modified_id); CREATE UNIQUE INDEX ix_master_other_unique_id_1 ON master USING btree (other_unique_id_1); DDL for child table CREATE TABLE child ( <date column> timestamp NULL DEFAULT now(), c_id serial4 NOT NULL, m_id int4 NULL, <other_column_1> varchar(50) NULL, <other_column_2> varchar(50) NULL, <other_column_3> bool NULL, <other_column_4> varchar(50) NULL, <other_column_5> varchar(2) NULL, <other_column_6> varchar(50) NULL, <other_column_7> varchar(10) NULL, <other_column_8> text NULL, CONSTRAINT child_pkey PRIMARY KEY (c_id), CONSTRAINT child_master_fkey FOREIGN KEY (m_id) REFERENCES master(m_id) ); CREATE INDEX ix_child_m_id ON child USING btree (m_id);
I think the root of the problem is this gross misestimate. -> Bitmap Index Scan on ix_master_modified_id (cost=0.00..40.71 rows=2153 width=0) (actual time=0.027..0.027 rows=8 loops=3) You said you already did VACUUM ANALYZE the table. In that case you might need to increase the stats size, and then ANALYZE again, to get a better estimate. You could either change the default_statistics_target globally, or target this one column with: alter table master alter modified_id set statistics 10000; analyze master; You might not need to increase it all the way to 10000 (the max allowed), but if you only do it for one column I see no reason trying to fine tune it. Just break out the big guns right away. If it works, then you can worry about fine tuning it.
Seemingly Random Delay in queries
This is a follow up to this issue I posted a while ago. I have the following code: SET work_mem = '16MB'; SELECT s.start_date, s.end_date, s.resources, s.activity_index, r.resource_id, sa.usedresourceset FROM rm_o_resource_usage_instance_splits_new s INNER JOIN rm_o_resource_usage r ON s.usage_id = r.id INNER JOIN scheduledactivities sa ON s.activity_index = sa.activity_index AND r.schedule_id = sa.solution_id and s.solution = sa.solution_id WHERE r.schedule_id = 10 ORDER BY r.resource_id, s.start_date When I run EXPLAIN (ANALYZE, BUFFERS) I get the following: Sort (cost=3724.02..3724.29 rows=105 width=89) (actual time=245.802..247.573 rows=22302 loops=1) Sort Key: r.resource_id, s.start_date Sort Method: quicksort Memory: 6692kB Buffers: shared hit=198702 read=5993 written=612 -> Nested Loop (cost=703.76..3720.50 rows=105 width=89) (actual time=1.898..164.741 rows=22302 loops=1) Buffers: shared hit=198702 read=5993 written=612 -> Hash Join (cost=703.34..3558.54 rows=105 width=101) (actual time=1.815..11.259 rows=22302 loops=1) Hash Cond: (s.usage_id = r.id) Buffers: shared hit=3 read=397 written=2 -> Bitmap Heap Scan on rm_o_resource_usage_instance_splits_new s (cost=690.61..3486.58 rows=22477 width=69) (actual time=1.782..5.820 rows=22302 loops=1) Recheck Cond: (solution = 10) Heap Blocks: exact=319 Buffers: shared hit=2 read=396 written=2 -> Bitmap Index Scan on rm_o_resource_usage_instance_splits_new_solution_idx (cost=0.00..685.00 rows=22477 width=0) (actual time=1.609..1.609 rows=22302 loops=1) Index Cond: (solution = 10) Buffers: shared hit=2 read=77 -> Hash (cost=12.66..12.66 rows=5 width=48) (actual time=0.023..0.023 rows=1 loops=1) Buckets: 1024 Batches: 1 Memory Usage: 9kB Buffers: shared hit=1 read=1 -> Bitmap Heap Scan on rm_o_resource_usage r (cost=4.19..12.66 rows=5 width=48) (actual time=0.020..0.020 rows=1 loops=1) Recheck Cond: (schedule_id = 10) Heap Blocks: exact=1 Buffers: shared hit=1 read=1 -> Bitmap Index Scan on rm_o_resource_usage_sched (cost=0.00..4.19 rows=5 width=0) (actual time=0.017..0.017 rows=1 loops=1) Index Cond: (schedule_id = 10) Buffers: shared read=1 -> Index Scan using scheduledactivities_activity_index_idx on scheduledactivities sa (cost=0.42..1.53 rows=1 width=16) (actual time=0.004..0.007 rows=1 loops=22302) Index Cond: (activity_index = s.activity_index) Filter: (solution_id = 10) Rows Removed by Filter: 5 Buffers: shared hit=198699 read=5596 written=610 Planning time: 7.070 ms Execution time: 248.691 ms Every time I run EXPLAIN, I get roughly the same results. The Execution Time is always between 170ms and 250ms, which, to me is perfectly fine. However, when this query is run through a C++ project (using PQexec(conn, query) where conn is a dedicated connection, and query is the above query), the time it takes seems to vary widely. In general, the query is very quick, and you don't notice a delay. The problem is, that on occasion, this query will take 2 to 3 minutes to complete. If I open the pgadmin, and have a look at the "server activity" for the database, there's about 30 or so connections, mostly sitting at "idle". The above query's connection is marked as "active", and will stay as "active" for several minutes. I am at a loss of why it randomly takes several minutes to complete the same query, with no change in data in the DB either. I have tried increasing the work_mem which didn't make any difference (nor did I really expect it to). Any help or suggestions would be greatly appreciated. There isn't any more specific tags, but I'm currently using Postgres 10.11, but it's also been an issue on other versions of 10.x. System is a Xeon quad-core # 3.4Ghz, with SSD and 24GB of memory. Per jjanes's suggestion, I put in the auto_explain. Eventually go this output: duration: 128057.373 ms plan: Query Text: SET work_mem = '32MB';SELECT s.start_date, s.end_date, s.resources, s.activity_index, r.resource_id, sa.usedresourceset FROM rm_o_resource_usage_instance_splits_new s INNER JOIN rm_o_resource_usage r ON s.usage_id = r.id INNER JOIN scheduledactivities sa ON s.activity_index = sa.activity_index AND r.schedule_id = sa.solution_id and s.solution = sa.solution_id WHERE r.schedule_id = 12642 ORDER BY r.resource_id, s.start_date Sort (cost=14.36..14.37 rows=1 width=98) (actual time=128042.083..128043.287 rows=21899 loops=1) Output: s.start_date, s.end_date, s.resources, s.activity_index, r.resource_id, sa.usedresourceset Sort Key: r.resource_id, s.start_date Sort Method: quicksort Memory: 6585kB Buffers: shared hit=21198435 read=388 dirtied=119 -> Nested Loop (cost=0.85..14.35 rows=1 width=98) (actual time=4.995..127958.935 rows=21899 loops=1) Output: s.start_date, s.end_date, s.resources, s.activity_index, r.resource_id, sa.usedresourceset Join Filter: (s.activity_index = sa.activity_index) Rows Removed by Join Filter: 705476285 Buffers: shared hit=21198435 read=388 dirtied=119 -> Nested Loop (cost=0.42..9.74 rows=1 width=110) (actual time=0.091..227.705 rows=21899 loops=1) Output: s.start_date, s.end_date, s.resources, s.activity_index, s.solution, r.resource_id, r.schedule_id Inner Unique: true Join Filter: (s.usage_id = r.id) Buffers: shared hit=22102 read=388 dirtied=119 -> Index Scan using rm_o_resource_usage_instance_splits_new_solution_idx on public.rm_o_resource_usage_instance_splits_new s (cost=0.42..8.44 rows=1 width=69) (actual time=0.082..17.418 rows=21899 loops=1) Output: s.start_time, s.end_time, s.resources, s.activity_index, s.usage_id, s.start_date, s.end_date, s.solution Index Cond: (s.solution = 12642) Buffers: shared hit=203 read=388 dirtied=119 -> Seq Scan on public.rm_o_resource_usage r (cost=0.00..1.29 rows=1 width=57) (actual time=0.002..0.002 rows=1 loops=21899) Output: r.id, r.schedule_id, r.resource_id Filter: (r.schedule_id = 12642) Rows Removed by Filter: 26 Buffers: shared hit=21899 -> Index Scan using scheduled_activities_idx on public.scheduledactivities sa (cost=0.42..4.60 rows=1 width=16) (actual time=0.006..4.612 rows=32216 loops=21899) Output: sa.usedresourceset, sa.activity_index, sa.solution_id Index Cond: (sa.solution_id = 12642) Buffers: shared hit=21176333",,,,,,,,,"" EDIT: Full definitions of the tables are below: CREATE TABLE public.rm_o_resource_usage_instance_splits_new ( start_time integer NOT NULL, end_time integer NOT NULL, resources jsonb NOT NULL, activity_index integer NOT NULL, usage_id bigint NOT NULL, start_date text COLLATE pg_catalog."default" NOT NULL, end_date text COLLATE pg_catalog."default" NOT NULL, solution bigint NOT NULL, CONSTRAINT rm_o_resource_usage_instance_splits_new_pkey PRIMARY KEY (start_time, activity_index, usage_id), CONSTRAINT rm_o_resource_usage_instance_splits_new_solution_fkey FOREIGN KEY (solution) REFERENCES public.rm_o_schedule_stats (id) MATCH SIMPLE ON UPDATE CASCADE ON DELETE CASCADE, CONSTRAINT rm_o_resource_usage_instance_splits_new_usage_id_fkey FOREIGN KEY (usage_id) REFERENCES public.rm_o_resource_usage (id) MATCH SIMPLE ON UPDATE CASCADE ON DELETE CASCADE ) WITH ( OIDS = FALSE ) TABLESPACE pg_default; CREATE INDEX rm_o_resource_usage_instance_splits_new_activity_idx ON public.rm_o_resource_usage_instance_splits_new USING btree (activity_index ASC NULLS LAST) TABLESPACE pg_default; CREATE INDEX rm_o_resource_usage_instance_splits_new_solution_idx ON public.rm_o_resource_usage_instance_splits_new USING btree (solution ASC NULLS LAST) TABLESPACE pg_default; CREATE INDEX rm_o_resource_usage_instance_splits_new_usage_idx ON public.rm_o_resource_usage_instance_splits_new USING btree (usage_id ASC NULLS LAST) TABLESPACE pg_default; CREATE TABLE public.rm_o_resource_usage ( id bigint NOT NULL DEFAULT nextval('rm_o_resource_usage_id_seq'::regclass), schedule_id bigint NOT NULL, resource_id text COLLATE pg_catalog."default" NOT NULL, CONSTRAINT rm_o_resource_usage_pkey PRIMARY KEY (id), CONSTRAINT rm_o_resource_usage_schedule_id_fkey FOREIGN KEY (schedule_id) REFERENCES public.rm_o_schedule_stats (id) MATCH SIMPLE ON UPDATE CASCADE ON DELETE CASCADE ) WITH ( OIDS = FALSE ) TABLESPACE pg_default; CREATE INDEX rm_o_resource_usage_idx ON public.rm_o_resource_usage USING btree (id ASC NULLS LAST) TABLESPACE pg_default; CREATE INDEX rm_o_resource_usage_sched ON public.rm_o_resource_usage USING btree (schedule_id ASC NULLS LAST) TABLESPACE pg_default; CREATE TABLE public.scheduledactivities ( id bigint NOT NULL DEFAULT nextval('scheduledactivities_id_seq'::regclass), solution_id bigint NOT NULL, activity_id text COLLATE pg_catalog."default" NOT NULL, sequence_index integer, startminute integer, finishminute integer, issue text COLLATE pg_catalog."default", activity_index integer NOT NULL, is_objective boolean NOT NULL, usedresourceset integer DEFAULT '-1'::integer, start timestamp without time zone, finish timestamp without time zone, is_ore boolean, is_ignored boolean, CONSTRAINT scheduled_activities_pkey PRIMARY KEY (id), CONSTRAINT scheduledactivities_solution_id_fkey FOREIGN KEY (solution_id) REFERENCES public.rm_o_schedule_stats (id) MATCH SIMPLE ON UPDATE CASCADE ON DELETE CASCADE ) WITH ( OIDS = FALSE ) TABLESPACE pg_default; CREATE INDEX scheduled_activities_activity_id_idx ON public.scheduledactivities USING btree (activity_id COLLATE pg_catalog."default" ASC NULLS LAST) TABLESPACE pg_default; CREATE INDEX scheduled_activities_id_idx ON public.scheduledactivities USING btree (id ASC NULLS LAST) TABLESPACE pg_default; CREATE INDEX scheduled_activities_idx ON public.scheduledactivities USING btree (solution_id ASC NULLS LAST) TABLESPACE pg_default; CREATE INDEX scheduledactivities_activity_index_idx ON public.scheduledactivities USING btree (activity_index ASC NULLS LAST) TABLESPACE pg_default; EDIT: Additional output from auto_explain after adding index on scheduledactivities (solution_id, activity_index) Output: s.start_date, s.end_date, s.resources, s.activity_index, r.resource_id, sa.usedresourceset Sort Key: r.resource_id, s.start_date Sort Method: quicksort Memory: 6283kB Buffers: shared hit=20159117 read=375 dirtied=190 -> Nested Loop (cost=0.85..10.76 rows=1 width=100) (actual time=5.518..122489.627 rows=20761 loops=1) Output: s.start_date, s.end_date, s.resources, s.activity_index, r.resource_id, sa.usedresourceset Join Filter: (s.activity_index = sa.activity_index) Rows Removed by Join Filter: 668815615 Buffers: shared hit=20159117 read=375 dirtied=190 -> Nested Loop (cost=0.42..5.80 rows=1 width=112) (actual time=0.057..217.563 rows=20761 loops=1) Output: s.start_date, s.end_date, s.resources, s.activity_index, s.solution, r.resource_id, r.schedule_id Inner Unique: true Join Filter: (s.usage_id = r.id) Buffers: shared hit=20947 read=375 dirtied=190 -> Index Scan using rm_o_resource_usage_instance_splits_new_solution_idx on public.rm_o_resource_usage_instance_splits_new s (cost=0.42..4.44 rows=1 width=69) (actual time=0.049..17.622 rows=20761 loops=1) Output: s.start_time, s.end_time, s.resources, s.activity_index, s.usage_id, s.start_date, s.end_date, s.solution Index Cond: (s.solution = 12644) Buffers: shared hit=186 read=375 dirtied=190 -> Seq Scan on public.rm_o_resource_usage r (cost=0.00..1.35 rows=1 width=59) (actual time=0.002..0.002 rows=1 loops=20761) Output: r.id, r.schedule_id, r.resource_id Filter: (r.schedule_id = 12644) Rows Removed by Filter: 22 Buffers: shared hit=20761 -> Index Scan using scheduled_activities_idx on public.scheduledactivities sa (cost=0.42..4.94 rows=1 width=16) (actual time=0.007..4.654 rows=32216 loops=20761) Output: sa.usedresourceset, sa.activity_index, sa.solution_id Index Cond: (sa.solution_id = 12644) Buffers: shared hit=20138170",,,,,,,,,"" The easiest way to reproduce the issue is to add more values to the three tables. I didn't delete any, only did a few thousand INSERTs.
-> Index Scan using .. s (cost=0.42..8.44 rows=1 width=69) (actual time=0.082..17.418 rows=21899 loops=1) Index Cond: (s.solution = 12642) The planner thinks it will find 1 row, and instead finds 21899. That error can pretty clearly lead to bad plans. And a single equality condition should be estimated quite accurately, so I'd say the statistics on your table are way off. It could be that the autovac launcher is tuned poorly so it doesn't run often enough, or it could be that select parts of your data change very rapidly (did you just insert 21899 rows with s.solution = 12642 immediately before running the query?) and so the stats can't be kept accurate enough. -> Nested Loop ... Join Filter: (s.activity_index = sa.activity_index) Rows Removed by Join Filter: 705476285 -> ... -> Index Scan using scheduled_activities_idx on public.scheduledactivities sa (cost=0.42..4.60 rows=1 width=16) (actual time=0.006..4.612 rows=32216 loops=21899) Output: sa.usedresourceset, sa.activity_index, sa.solution_id Index Cond: (sa.solution_id = 12642) If you can't get it to use the Hash Join, you can at least reduce the harm of the Nested Loop by building an index on scheduledactivities (solution_id, activity_index). That way the activity_index criterion could be part of the Index Condition, rather than being a Join Filter. You could probably then drop the index exclusively on solution_id, as there is little point in maintaining both indexes.
The SQL statement of the fast plan is using WHERE r.schedule_id = 10 and returns about 22000 rows (with estimated 105). The SQL statement of the slow plan is using WHERE r.schedule_id = 12642 and returns about 21000 rows (with estimated only 1). The slow plan is using nested loops instead of hash joins: maybe because there is a bad estimation for joins: estimated rows is 1 but actual rows is 21899. For example in this step: Nested Loop (cost=0.42..9.74 rows=1 width=110) (actual time=0.091..227.705 rows=21899 loops=1) If data does not change there is maybe a statistic issue (skew data) for some columns.
PostgreSQL: improve the performance when counting the distinct
I am currently working on improving the performance of our db. And I need some help from you. I have a table and its index like this CREATE TABLE public.ar ( id integer NOT NULL DEFAULT nextval('id_seq'::regclass), user_id integer NOT NULL, duration double precision, is_idle boolean NOT NULL, activity_id integer NOT NULL, device_id integer NOT NULL, calendar_id integer, on_taskness integer, week_id integer, some_other_column_below, CONSTRAINT id_ PRIMARY KEY (id), CONSTRAINT a_unique_key UNIQUE (user_id, device_id, start_time_local, start_time_utc, end_time_local, end_time_utc) ) CREATE INDEX ar_idx ON public.ar USING btree (week_id, calendar_id, user_id, activity_id, duration, on_taskness, is_idle) TABLESPACE pg_default; Then I am trying to run a query like this EXPLAIN ANALYZE SELECT COUNT(*) FROM ( SELECT ar.user_id FROM ar WHERE ar.user_id = ANY(array[some_data]) -- data size is 352 AND ROUND(ar.duration) >0 AND ar.is_idle = false AND ar.week_id = ANY(ARRAY[some_data]) -- data size is 37 AND ar.calendar_id = ANY(array[some_data]) -- data size is 16716 GROUP by ar.user_id ) tmp; And below is the explain result Aggregate (cost=31389954.72..31389954.73 rows=1 width=8) (actual time=252020.695..252020.695 rows=1 loops=1) -> Group (cost=31389032.69..31389922.37 rows=2588 width=4) (actual time=251089.270..252020.659 rows=351 loops=1) Group Key: ar.user_id -> Sort (cost=31389032.69..31389477.53 rows=177935 width=4) (actual time=251089.268..251776.202 rows=6993358 loops=1) Sort Key: ar.user_id Sort Method: external merge Disk: 95672kB -> Bitmap Heap Scan on ar (cost=609015.18..31371079.88 rows=177935 width=4) (actual time=1670.413..248939.440 rows=6993358 loops=1) Recheck Cond: ((week_id = ANY ('{some_data}'::integer[])) AND (user_id = ANY ('{some_data}'::integer[]))) Rows Removed by Index Recheck: 2081028 Filter: ((NOT is_idle) AND (round(duration) > '0'::double precision) AND (calendar_id = ANY ('{some_data}'::integer[]))) Rows Removed by Filter: 534017 Heap Blocks: exact=29551 lossy=313127 -> BitmapAnd (cost=609015.18..609015.18 rows=1357521 width=0) (actual time=1666.334..1666.334 rows=0 loops=1) -> Bitmap Index Scan on test_index_only_scan_idx (cost=0.00..272396.77 rows=6970353 width=0) (actual time=614.366..614.366 rows=7269830 loops=1) Index Cond: ((week_id = ANY ('{some_data}'::integer[])) AND (is_idle = false)) -> Bitmap Index Scan on unique_key (cost=0.00..336529.20 rows=9948573 width=0) (actual time=1041.999..1041.999 rows=14959355 loops=1) Index Cond: (user_id = ANY ('{some_data}'::integer[])) Planning time: 25.563 ms Execution time: 252029.237 ms I used distinct as well, and the result is the same. So my questions are below. The ar_idx contains user_id, but when searching for rows, why does it use the unique_key instead of the index I created? I thought group by will not do the sort(that is why I did not choose distinct), but why does the sort happen in the explain analyze? The running time is pretty long(more than 4 minutes). How do I make it faster? Is the index wrong? Or anything else I can do. Be advised, the ar table contains 51585203 rows. Any help will be appreciated. Thx. ---------------------------update-------------------------- After I created this index, everything goes really fast now. I don't understand why, anyone can explain this to me? CREATE INDEX ar_1_idx ON public.ar USING btree (calendar_id, user_id) TABLESPACE pg_default; And I changed the old index to CREATE INDEX ar_idx ON public.ar USING btree (week_id, calendar, user_id, activity_id, duration, on_taskness, start_time_local, end_time_local) WHERE is_idle IS FALSE TABLESPACE pg_default; -----updated analyze results----------- Aggregate (cost=31216435.97..31216435.98 rows=1 width=8) (actual time=13206.941..13206.941 rows=1 loops=1) Buffers: shared hit=25940518 read=430315, temp read=31079 written=31079 -> Group (cost=31215436.80..31216403.88 rows=2567 width=4) (actual time=12239.336..13206.894 rows=351 loops=1) Group Key: ar.user_id Buffers: shared hit=25940518 read=430315, temp read=31079 written=31079 -> Sort (cost=31215436.80..31215920.34 rows=193417 width=4) (actual time=12239.334..12932.801 rows=6993358 loops=1) Sort Key: ar.user_id Sort Method: external merge Disk: 95664kB Buffers: shared hit=25940518 read=430315, temp read=31079 written=31079 -> Index Scan using ar_1_idx on activity_report ar (cost=0.56..31195807.48 rows=193417 width=4) (actual time=0.275..10387.051 rows=6993358 loops=1) Index Cond: ((calendar_id = ANY ('{some_data}'::integer[])) AND (user_id = ANY ('{some_data}'::integer[]))) Filter: ((NOT is_idle) AND (round(duration) > '0'::double precision) AND (week_id = ANY ('{some_data}'::integer[]))) Rows Removed by Filter: 590705 Buffers: shared hit=25940518 read=430315 Planning time: 25.577 ms Execution time: 13217.611 ms
postgresql doesn't use index for primary key = foreign key
I have 3 main tables, ts_entity(id,short_name,name,type_id) ts_entry_entity(id,entity_id,entry_id) ts_entry(id, ... other columns ...) All the id columns are UUID, and have a Btree index. ts_entry_entity.entity_id has foreign key to ts_entity.id, and also has Btree index. ts_entry_entity.entry_id also is foreign key, and also has Btree index. I have one SQL, like select ts_entity.id,ts_entity.short_name,ts_entity.name,ts_entry.id, ... ts_entry.otherColumns ... from ts_entity,ts_entry_entity,ts_entry where ts_entity.id=ts_entry_entity.entity_id and ts_entry_entity.entry_id=ts_entry.id and ... ts_entry.otherColumns='xxx' ... order by ts_entity.short_name limit 100 offset 0 And here comes the weird thing, "ts_entry_entity.entity_id=ts_entity.id" doesn't use any indexes and it costs about 50s. There is no where condition on table ts_entity. My question: Why does ts_entry_entity.entity_id=ts_entity.id not use an index? Why does it cost so much time? How can I optimize the SQL? Below is the explain analyze result. Limit (cost=235455.31..235455.41 rows=1 width=1808) (actual time=54590.304..54590.781 rows=100 loops=1) -> Unique (cost=235455.31..235455.41 rows=1 width=1808) (actual time=54590.301..54590.666 rows=100 loops=1) -> Sort (cost=235455.31..235455.32 rows=1 width=1808) (actual time=54590.297..54590.410 rows=100 loops=1) Sort Key: ts_entity.short_name, ts_entity.id, ts_entity.name, ts_entry_version.display_date, ts_entry.id, (formatdate(totimestamp(ts_entry_version.display_date, '-5'::character varying), 'MM/DD/YYYY'::charac ter varying)), ts_entry_version.submitted_date, (formatdate(totimestamp(ts_entry_version.submitted_date, '-5'::character varying), 'MM/DD/YYYY'::character varying)), ts_entry_type.name, (get_priority((ts_entry_version.prio rity)::integer)), ts_entry_version.priority, (get_sentiment((ts_entry_version.sentiment)::integer)), ts_entry_version.sentiment, (getdisplayvalue((ts_entry_version.source_id)::character varying, 0, ', '::character varying) ), ts_entry_version.source_id, (NULLIF((ts_entry_version.title)::text, ''::text)), ts_entry.submitted_date, (formatdate(totimestamp(ts_entry.submitted_date, '-5'::character varying), 'MM/DD/YYYY'::character varying)), (get displayvalue((ts_entry_version.submitter_id)::character varying, 0, ', '::character varying)), ts_entry_version.submitter_id, entryadhoc_o9e2c9f871634dd3aeafe9bdced2e34f.owner_id, (getdisplayvalue(toentityid((entryadhoc_o9 e2c9f871634dd3aeafe9bdced2e34f.value)::character varying, '23f03fe70a16aed0d7e210357164e401'::character varying), 0, ', '::character varying)), (toentityid((entryadhoc_o9e2c9f871634dd3aeafe9bdced2e34f.value)::character var ying, '23f03fe70a16aed0d7e210357164e401'::character varying)), entryadhoc_td66ad96a9ab472db3cf1279b65baa69.owner_id, (totimestamp((entryadhoc_td66ad96a9ab472db3cf1279b65baa69.value)::character varying, '-5'::character vary ing)), (formatdate(totimestamp((entryadhoc_td66ad96a9ab472db3cf1279b65baa69.value)::character varying, '-5'::character varying), 'MM/DD/YYYY'::character varying)), entryadhoc_z3757638d8d64373ad835c3523a6a70b.owner_id, (tot imestamp((entryadhoc_z3757638d8d64373ad835c3523a6a70b.value)::character varying, '-5'::character varying)), (formatdate(totimestamp((entryadhoc_z3757638d8d64373ad835c3523a6a70b.value)::character varying, '-5'::character va rying), 'MM/DD/YYYY'::character varying)), entryadhoc_i0f819c1244b427794a83767eaa68e73.owner_id, (totimestamp((entryadhoc_i0f819c1244b427794a83767eaa68e73.value)::character varying, '-5'::character varying)), (formatdate(t otimestamp((entryadhoc_i0f819c1244b427794a83767eaa68e73.value)::character varying, '-5'::character varying), 'MM/DD/YYYY'::character varying)), entryadhoc_i7f5d5035cac421daa9879c1e21ec63f.owner_id, (getdisplayvalue(toentit yid((entryadhoc_i7f5d5035cac421daa9879c1e21ec63f.value)::character varying, '23f03fe70a16aed0d7e210357164e401'::character varying), 0, ', '::character varying)), (toentityid((entryadhoc_i7f5d5035cac421daa9879c1e21ec63f.val ue)::character varying, '23f03fe70a16aed0d7e210357164e401'::character varying)), entryadhoc_v7f9c1146ee24742a73b83526dc66df7.owner_id, (NULLIF(entryadhoc_v7f9c1146ee24742a73b83526dc66df7.value, ''::text)) Sort Method: external merge Disk: 3360kB -> Nested Loop (cost=22979.01..235455.30 rows=1 width=1808) (actual time=94.889..54532.919 rows=2846 loops=1) Join Filter: (ts_entry_entity.entity_id = ts_entity.id) Rows Removed by Join Filter: 34363583 -> Nested Loop (cost=22979.01..234676.15 rows=1 width=987) (actual time=78.801..2914.864 rows=2846 loops=1) -> Nested Loop Anti Join (cost=22978.59..234675.43 rows=1 width=987) (actual time=78.776..2867.254 rows=2846 loops=1) -> Hash Join (cost=22978.17..63457.52 rows=258 width=987) (actual time=78.614..2573.586 rows=2846 loops=1) Hash Cond: (ts_entry.current_version_id = ts_entry_version.id) -> Hash Left Join (cost=19831.38..59727.56 rows=154823 width=383) (actual time=47.558..2391.088 rows=155061 loops=1) Hash Cond: (ts_entry.id = entryadhoc_v7f9c1146ee24742a73b83526dc66df7.owner_id) -> Hash Left Join (cost=16526.15..54467.69 rows=154823 width=337) (actual time=38.534..2138.354 rows=155061 loops=1) Hash Cond: (ts_entry.id = entryadhoc_i7f5d5035cac421daa9879c1e21ec63f.owner_id) -> Hash Left Join (cost=13220.92..49207.82 rows=154823 width=291) (actual time=30.462..1888.735 rows=155061 loops=1) Hash Cond: (ts_entry.id = entryadhoc_i0f819c1244b427794a83767eaa68e73.owner_id) -> Hash Left Join (cost=9915.69..43947.95 rows=154823 width=245) (actual time=22.268..1640.688 rows=155061 loops=1) Hash Cond: (ts_entry.id = entryadhoc_z3757638d8d64373ad835c3523a6a70b.owner_id) -> Hash Left Join (cost=6610.46..38688.08 rows=154823 width=199) (actual time=19.612..1409.457 rows=155061 loops=1) Hash Cond: (ts_entry.id = entryadhoc_td66ad96a9ab472db3cf1279b65baa69.owner_id) -> Hash Left Join (cost=3305.23..33428.21 rows=154823 width=153) (actual time=12.431..1161.689 rows=155061 loops=1) Hash Cond: (ts_entry.id = entryadhoc_o9e2c9f871634dd3aeafe9bdced2e34f.owner_id) -> Seq Scan on ts_entry (cost=0.00..28168.34 rows=154823 width=107) (actual time=0.101..898.818 rows=155061 loops=1) Filter: ((NOT is_draft) AND (class <> 2)) Rows Removed by Filter: 236596 -> Hash (cost=3292.29..3292.29 rows=1035 width=46) (actual time=12.304..12.304 rows=2846 loops=1) Buckets: 4096 (originally 2048) Batches: 1 (originally 1) Memory Usage: 305kB -> Bitmap Heap Scan on ts_attribute entryadhoc_o9e2c9f871634dd3aeafe9bdced2e34f (cost=40.45..3292.29 rows=1035 width=46) (actual time=1.191 ..9.030 rows=2846 loops=1) Recheck Cond: (def_id = 'b4e9878722eb409c9fdfff3fdba582a3'::bpchar) More details about tables: ts_entity(id,short_name,name,type_id) ts_entry_entity(id,entity_id,entry_id) ts_entry(id,version_id) ts_entry_version(id,entry_id,submitted_date,title,submitter) ts_attribute(id,attribute_definition_id,entry_id,value) ts_attribute_definition(id,name) As you can see, ts_entry_version will save all versions for one entry. ts_attribute is used for the extendable columns for entry. More details about the SQL We have several filters on ts_entry_version columns and ts_attribute.value. ts_attribute.value is varchar, but the content may be time milliseconds, normal string value, one or serval id values. The structure of the SQL is like below: select ts_entity.short_name, ts_entry_version.title, ts_attribute.value from ts_entity, ts_entry_entity,ts_entry left join ts_attribute on ts_entry.id=ts_attribute.entry_id and ts_attribute.attribute_definition_id='xxx' where ts_entity.id=ts_entry_entity.entity_id and ts_entry_entity.entry_id=ts_entry.id and ts_entry.version_id=ts_entry_version.id and ts_entry_version.title like '%xxx%' order by ts_entity.short_name asc limit 100 offset 0
I found clue in PostgreSQL official documentation, https://www.postgresql.org/docs/current/static/runtime-config-query.html Change the configuration and the Query Optimizer will prefer to use indexes.
Optimize query, PostgreSQL
How can i optimize this query? I want to find all rows from firms2branches by project_id that exists in firms and accounts_premium. My tables: -- 50000 rows CREATE TABLE firms ( id bigserial NOT NULL, firm_id bigint NOT NULL, CONSTRAINT firms_pkey PRIMARY KEY (id) ) -- 2 300 000 rows CREATE TABLE firms2branches ( firm_id bigint NOT NULL, branch_id bigint NOT NULL, project_id bigint NOT NULL ) CREATE INDEX firms2branches_firm_id_idx ON firms2branches USING btree(firm_id); -- 6500 rows CREATE TABLE accounts_premium ( firm_id bigint NOT NULL, is_active boolean NOT NULL DEFAULT false, CONSTRAINT accounts_premium_pkey PRIMARY KEY (firm_id) ) CREATE INDEX accounts_premium_is_active_idx ON accounts_premium USING btree(is_active); Query (with cold cache): EXPLAIN (ANALYZE) SELECT firms2branches.branch_id, firms2branches.firm_id FROM firms2branches JOIN firms ON firms.firm_id = firms2branches.firm_id JOIN accounts_premium ON accounts_premium.firm_id = firms.id AND accounts_premium.is_active = TRUE WHERE firms2branches.project_id = 21 Result (https://explain.depesz.com/s/oVNH): Nested Loop (cost=22.12..6958.10 rows=355 width=16) (actual time=151.123..417.764 rows=31 loops=1) Buffers: shared hit=7176 read=3371 -> Nested Loop (cost=21.69..3100.40 rows=1435 width=8) (actual time=0.905..58.314 rows=1378 loops=1) Buffers: shared hit=3250 read=961 -> Bitmap Heap Scan on accounts_premium (cost=21.40..226.90 rows=1435 width=8) (actual time=0.615..1.211 rows=1378 loops=1) Filter: is_active Heap Blocks: exact=61 Buffers: shared hit=61 read=6 -> Bitmap Index Scan on accounts_premium_is_active_idx (cost=0.00..21.04 rows=1435 width=0) (actual time=0.594..0.594 rows=1435 loops=1) Index Cond: (is_active = true) Buffers: shared read=6 -> Index Scan using firms_pkey on firms (cost=0.29..1.90 rows=1 width=16) (actual time=0.040..0.041 rows=1 loops=1378) Index Cond: (id = accounts_premium.firm_id) Buffers: shared hit=3189 read=955 -> Index Scan using firms2branches_firm_id_idx on firms2branches (cost=0.43..2.59 rows=1 width=16) (actual time=0.259..0.260 rows=0 loops=1378) Index Cond: (firm_id = firms.firm_id) Filter: (project_id = 21::bigint) Rows Removed by Filter: 2 Buffers: shared hit=3926 read=2410 Planning time: 6.164 ms Execution time: 417.843 ms