]> gcc.gnu.org Git - gcc.git/blob - gcc/tree-sra.c
hash-table.h: Update comments.
[gcc.git] / gcc / tree-sra.c
1 /* Scalar Replacement of Aggregates (SRA) converts some structure
2 references into scalar references, exposing them to the scalar
3 optimizers.
4 Copyright (C) 2008-2015 Free Software Foundation, Inc.
5 Contributed by Martin Jambor <mjambor@suse.cz>
6
7 This file is part of GCC.
8
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
13
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
22
23 /* This file implements Scalar Reduction of Aggregates (SRA). SRA is run
24 twice, once in the early stages of compilation (early SRA) and once in the
25 late stages (late SRA). The aim of both is to turn references to scalar
26 parts of aggregates into uses of independent scalar variables.
27
28 The two passes are nearly identical, the only difference is that early SRA
29 does not scalarize unions which are used as the result in a GIMPLE_RETURN
30 statement because together with inlining this can lead to weird type
31 conversions.
32
33 Both passes operate in four stages:
34
35 1. The declarations that have properties which make them candidates for
36 scalarization are identified in function find_var_candidates(). The
37 candidates are stored in candidate_bitmap.
38
39 2. The function body is scanned. In the process, declarations which are
40 used in a manner that prevent their scalarization are removed from the
41 candidate bitmap. More importantly, for every access into an aggregate,
42 an access structure (struct access) is created by create_access() and
43 stored in a vector associated with the aggregate. Among other
44 information, the aggregate declaration, the offset and size of the access
45 and its type are stored in the structure.
46
47 On a related note, assign_link structures are created for every assign
48 statement between candidate aggregates and attached to the related
49 accesses.
50
51 3. The vectors of accesses are analyzed. They are first sorted according to
52 their offset and size and then scanned for partially overlapping accesses
53 (i.e. those which overlap but one is not entirely within another). Such
54 an access disqualifies the whole aggregate from being scalarized.
55
56 If there is no such inhibiting overlap, a representative access structure
57 is chosen for every unique combination of offset and size. Afterwards,
58 the pass builds a set of trees from these structures, in which children
59 of an access are within their parent (in terms of offset and size).
60
61 Then accesses are propagated whenever possible (i.e. in cases when it
62 does not create a partially overlapping access) across assign_links from
63 the right hand side to the left hand side.
64
65 Then the set of trees for each declaration is traversed again and those
66 accesses which should be replaced by a scalar are identified.
67
68 4. The function is traversed again, and for every reference into an
69 aggregate that has some component which is about to be scalarized,
70 statements are amended and new statements are created as necessary.
71 Finally, if a parameter got scalarized, the scalar replacements are
72 initialized with values from respective parameter aggregates. */
73
74 #include "config.h"
75 #include "system.h"
76 #include "coretypes.h"
77 #include "alloc-pool.h"
78 #include "tm.h"
79 #include "alias.h"
80 #include "symtab.h"
81 #include "tree.h"
82 #include "fold-const.h"
83 #include "predict.h"
84 #include "hard-reg-set.h"
85 #include "function.h"
86 #include "dominance.h"
87 #include "cfg.h"
88 #include "basic-block.h"
89 #include "tree-ssa-alias.h"
90 #include "internal-fn.h"
91 #include "tree-eh.h"
92 #include "gimple-expr.h"
93 #include "gimple.h"
94 #include "stor-layout.h"
95 #include "gimplify.h"
96 #include "gimple-iterator.h"
97 #include "gimplify-me.h"
98 #include "gimple-walk.h"
99 #include "bitmap.h"
100 #include "gimple-ssa.h"
101 #include "tree-cfg.h"
102 #include "tree-phinodes.h"
103 #include "ssa-iterators.h"
104 #include "stringpool.h"
105 #include "tree-ssanames.h"
106 #include "rtl.h"
107 #include "flags.h"
108 #include "insn-config.h"
109 #include "expmed.h"
110 #include "dojump.h"
111 #include "explow.h"
112 #include "calls.h"
113 #include "emit-rtl.h"
114 #include "varasm.h"
115 #include "stmt.h"
116 #include "expr.h"
117 #include "tree-dfa.h"
118 #include "tree-ssa.h"
119 #include "tree-pass.h"
120 #include "plugin-api.h"
121 #include "ipa-ref.h"
122 #include "cgraph.h"
123 #include "symbol-summary.h"
124 #include "ipa-prop.h"
125 #include "params.h"
126 #include "target.h"
127 #include "dbgcnt.h"
128 #include "tree-inline.h"
129 #include "gimple-pretty-print.h"
130 #include "ipa-inline.h"
131 #include "ipa-utils.h"
132 #include "builtins.h"
133
134 /* Enumeration of all aggregate reductions we can do. */
135 enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
136 SRA_MODE_EARLY_INTRA, /* early intraprocedural SRA */
137 SRA_MODE_INTRA }; /* late intraprocedural SRA */
138
139 /* Global variable describing which aggregate reduction we are performing at
140 the moment. */
141 static enum sra_mode sra_mode;
142
143 struct assign_link;
144
145 /* ACCESS represents each access to an aggregate variable (as a whole or a
146 part). It can also represent a group of accesses that refer to exactly the
147 same fragment of an aggregate (i.e. those that have exactly the same offset
148 and size). Such representatives for a single aggregate, once determined,
149 are linked in a linked list and have the group fields set.
150
151 Moreover, when doing intraprocedural SRA, a tree is built from those
152 representatives (by the means of first_child and next_sibling pointers), in
153 which all items in a subtree are "within" the root, i.e. their offset is
154 greater or equal to offset of the root and offset+size is smaller or equal
155 to offset+size of the root. Children of an access are sorted by offset.
156
157 Note that accesses to parts of vector and complex number types always
158 represented by an access to the whole complex number or a vector. It is a
159 duty of the modifying functions to replace them appropriately. */
160
161 struct access
162 {
163 /* Values returned by `get_ref_base_and_extent' for each component reference
164 If EXPR isn't a component reference just set `BASE = EXPR', `OFFSET = 0',
165 `SIZE = TREE_SIZE (TREE_TYPE (expr))'. */
166 HOST_WIDE_INT offset;
167 HOST_WIDE_INT size;
168 tree base;
169
170 /* Expression. It is context dependent so do not use it to create new
171 expressions to access the original aggregate. See PR 42154 for a
172 testcase. */
173 tree expr;
174 /* Type. */
175 tree type;
176
177 /* The statement this access belongs to. */
178 gimple stmt;
179
180 /* Next group representative for this aggregate. */
181 struct access *next_grp;
182
183 /* Pointer to the group representative. Pointer to itself if the struct is
184 the representative. */
185 struct access *group_representative;
186
187 /* If this access has any children (in terms of the definition above), this
188 points to the first one. */
189 struct access *first_child;
190
191 /* In intraprocedural SRA, pointer to the next sibling in the access tree as
192 described above. In IPA-SRA this is a pointer to the next access
193 belonging to the same group (having the same representative). */
194 struct access *next_sibling;
195
196 /* Pointers to the first and last element in the linked list of assign
197 links. */
198 struct assign_link *first_link, *last_link;
199
200 /* Pointer to the next access in the work queue. */
201 struct access *next_queued;
202
203 /* Replacement variable for this access "region." Never to be accessed
204 directly, always only by the means of get_access_replacement() and only
205 when grp_to_be_replaced flag is set. */
206 tree replacement_decl;
207
208 /* Is this particular access write access? */
209 unsigned write : 1;
210
211 /* Is this access an access to a non-addressable field? */
212 unsigned non_addressable : 1;
213
214 /* Is this access currently in the work queue? */
215 unsigned grp_queued : 1;
216
217 /* Does this group contain a write access? This flag is propagated down the
218 access tree. */
219 unsigned grp_write : 1;
220
221 /* Does this group contain a read access? This flag is propagated down the
222 access tree. */
223 unsigned grp_read : 1;
224
225 /* Does this group contain a read access that comes from an assignment
226 statement? This flag is propagated down the access tree. */
227 unsigned grp_assignment_read : 1;
228
229 /* Does this group contain a write access that comes from an assignment
230 statement? This flag is propagated down the access tree. */
231 unsigned grp_assignment_write : 1;
232
233 /* Does this group contain a read access through a scalar type? This flag is
234 not propagated in the access tree in any direction. */
235 unsigned grp_scalar_read : 1;
236
237 /* Does this group contain a write access through a scalar type? This flag
238 is not propagated in the access tree in any direction. */
239 unsigned grp_scalar_write : 1;
240
241 /* Is this access an artificial one created to scalarize some record
242 entirely? */
243 unsigned grp_total_scalarization : 1;
244
245 /* Other passes of the analysis use this bit to make function
246 analyze_access_subtree create scalar replacements for this group if
247 possible. */
248 unsigned grp_hint : 1;
249
250 /* Is the subtree rooted in this access fully covered by scalar
251 replacements? */
252 unsigned grp_covered : 1;
253
254 /* If set to true, this access and all below it in an access tree must not be
255 scalarized. */
256 unsigned grp_unscalarizable_region : 1;
257
258 /* Whether data have been written to parts of the aggregate covered by this
259 access which is not to be scalarized. This flag is propagated up in the
260 access tree. */
261 unsigned grp_unscalarized_data : 1;
262
263 /* Does this access and/or group contain a write access through a
264 BIT_FIELD_REF? */
265 unsigned grp_partial_lhs : 1;
266
267 /* Set when a scalar replacement should be created for this variable. */
268 unsigned grp_to_be_replaced : 1;
269
270 /* Set when we want a replacement for the sole purpose of having it in
271 generated debug statements. */
272 unsigned grp_to_be_debug_replaced : 1;
273
274 /* Should TREE_NO_WARNING of a replacement be set? */
275 unsigned grp_no_warning : 1;
276
277 /* Is it possible that the group refers to data which might be (directly or
278 otherwise) modified? */
279 unsigned grp_maybe_modified : 1;
280
281 /* Set when this is a representative of a pointer to scalar (i.e. by
282 reference) parameter which we consider for turning into a plain scalar
283 (i.e. a by value parameter). */
284 unsigned grp_scalar_ptr : 1;
285
286 /* Set when we discover that this pointer is not safe to dereference in the
287 caller. */
288 unsigned grp_not_necessarilly_dereferenced : 1;
289
290 /* Pool allocation new operator. */
291 inline void *operator new (size_t)
292 {
293 return pool.allocate ();
294 }
295
296 /* Delete operator utilizing pool allocation. */
297 inline void operator delete (void *ptr)
298 {
299 pool.remove ((access *) ptr);
300 }
301
302 /* Memory allocation pool. */
303 static pool_allocator<access> pool;
304 };
305
306 typedef struct access *access_p;
307
308
309 /* Alloc pool for allocating access structures. */
310 pool_allocator<struct access> access::pool ("SRA accesses", 16);
311
312 /* A structure linking lhs and rhs accesses from an aggregate assignment. They
313 are used to propagate subaccesses from rhs to lhs as long as they don't
314 conflict with what is already there. */
315 struct assign_link
316 {
317 struct access *lacc, *racc;
318 struct assign_link *next;
319
320 /* Pool allocation new operator. */
321 inline void *operator new (size_t)
322 {
323 return pool.allocate ();
324 }
325
326 /* Delete operator utilizing pool allocation. */
327 inline void operator delete (void *ptr)
328 {
329 pool.remove ((assign_link *) ptr);
330 }
331
332 /* Memory allocation pool. */
333 static pool_allocator<assign_link> pool;
334 };
335
336 /* Alloc pool for allocating assign link structures. */
337 pool_allocator<assign_link> assign_link::pool ("SRA links", 16);
338
339 /* Base (tree) -> Vector (vec<access_p> *) map. */
340 static hash_map<tree, auto_vec<access_p> > *base_access_vec;
341
342 /* Candidate hash table helpers. */
343
344 struct uid_decl_hasher : nofree_ptr_hash <tree_node>
345 {
346 static inline hashval_t hash (const tree_node *);
347 static inline bool equal (const tree_node *, const tree_node *);
348 };
349
350 /* Hash a tree in a uid_decl_map. */
351
352 inline hashval_t
353 uid_decl_hasher::hash (const tree_node *item)
354 {
355 return item->decl_minimal.uid;
356 }
357
358 /* Return true if the DECL_UID in both trees are equal. */
359
360 inline bool
361 uid_decl_hasher::equal (const tree_node *a, const tree_node *b)
362 {
363 return (a->decl_minimal.uid == b->decl_minimal.uid);
364 }
365
366 /* Set of candidates. */
367 static bitmap candidate_bitmap;
368 static hash_table<uid_decl_hasher> *candidates;
369
370 /* For a candidate UID return the candidates decl. */
371
372 static inline tree
373 candidate (unsigned uid)
374 {
375 tree_node t;
376 t.decl_minimal.uid = uid;
377 return candidates->find_with_hash (&t, static_cast <hashval_t> (uid));
378 }
379
380 /* Bitmap of candidates which we should try to entirely scalarize away and
381 those which cannot be (because they are and need be used as a whole). */
382 static bitmap should_scalarize_away_bitmap, cannot_scalarize_away_bitmap;
383
384 /* Obstack for creation of fancy names. */
385 static struct obstack name_obstack;
386
387 /* Head of a linked list of accesses that need to have its subaccesses
388 propagated to their assignment counterparts. */
389 static struct access *work_queue_head;
390
391 /* Number of parameters of the analyzed function when doing early ipa SRA. */
392 static int func_param_count;
393
394 /* scan_function sets the following to true if it encounters a call to
395 __builtin_apply_args. */
396 static bool encountered_apply_args;
397
398 /* Set by scan_function when it finds a recursive call. */
399 static bool encountered_recursive_call;
400
401 /* Set by scan_function when it finds a recursive call with less actual
402 arguments than formal parameters.. */
403 static bool encountered_unchangable_recursive_call;
404
405 /* This is a table in which for each basic block and parameter there is a
406 distance (offset + size) in that parameter which is dereferenced and
407 accessed in that BB. */
408 static HOST_WIDE_INT *bb_dereferences;
409 /* Bitmap of BBs that can cause the function to "stop" progressing by
410 returning, throwing externally, looping infinitely or calling a function
411 which might abort etc.. */
412 static bitmap final_bbs;
413
414 /* Representative of no accesses at all. */
415 static struct access no_accesses_representant;
416
417 /* Predicate to test the special value. */
418
419 static inline bool
420 no_accesses_p (struct access *access)
421 {
422 return access == &no_accesses_representant;
423 }
424
425 /* Dump contents of ACCESS to file F in a human friendly way. If GRP is true,
426 representative fields are dumped, otherwise those which only describe the
427 individual access are. */
428
429 static struct
430 {
431 /* Number of processed aggregates is readily available in
432 analyze_all_variable_accesses and so is not stored here. */
433
434 /* Number of created scalar replacements. */
435 int replacements;
436
437 /* Number of times sra_modify_expr or sra_modify_assign themselves changed an
438 expression. */
439 int exprs;
440
441 /* Number of statements created by generate_subtree_copies. */
442 int subtree_copies;
443
444 /* Number of statements created by load_assign_lhs_subreplacements. */
445 int subreplacements;
446
447 /* Number of times sra_modify_assign has deleted a statement. */
448 int deleted;
449
450 /* Number of times sra_modify_assign has to deal with subaccesses of LHS and
451 RHS reparately due to type conversions or nonexistent matching
452 references. */
453 int separate_lhs_rhs_handling;
454
455 /* Number of parameters that were removed because they were unused. */
456 int deleted_unused_parameters;
457
458 /* Number of scalars passed as parameters by reference that have been
459 converted to be passed by value. */
460 int scalar_by_ref_to_by_val;
461
462 /* Number of aggregate parameters that were replaced by one or more of their
463 components. */
464 int aggregate_params_reduced;
465
466 /* Numbber of components created when splitting aggregate parameters. */
467 int param_reductions_created;
468 } sra_stats;
469
470 static void
471 dump_access (FILE *f, struct access *access, bool grp)
472 {
473 fprintf (f, "access { ");
474 fprintf (f, "base = (%d)'", DECL_UID (access->base));
475 print_generic_expr (f, access->base, 0);
476 fprintf (f, "', offset = " HOST_WIDE_INT_PRINT_DEC, access->offset);
477 fprintf (f, ", size = " HOST_WIDE_INT_PRINT_DEC, access->size);
478 fprintf (f, ", expr = ");
479 print_generic_expr (f, access->expr, 0);
480 fprintf (f, ", type = ");
481 print_generic_expr (f, access->type, 0);
482 if (grp)
483 fprintf (f, ", grp_read = %d, grp_write = %d, grp_assignment_read = %d, "
484 "grp_assignment_write = %d, grp_scalar_read = %d, "
485 "grp_scalar_write = %d, grp_total_scalarization = %d, "
486 "grp_hint = %d, grp_covered = %d, "
487 "grp_unscalarizable_region = %d, grp_unscalarized_data = %d, "
488 "grp_partial_lhs = %d, grp_to_be_replaced = %d, "
489 "grp_to_be_debug_replaced = %d, grp_maybe_modified = %d, "
490 "grp_not_necessarilly_dereferenced = %d\n",
491 access->grp_read, access->grp_write, access->grp_assignment_read,
492 access->grp_assignment_write, access->grp_scalar_read,
493 access->grp_scalar_write, access->grp_total_scalarization,
494 access->grp_hint, access->grp_covered,
495 access->grp_unscalarizable_region, access->grp_unscalarized_data,
496 access->grp_partial_lhs, access->grp_to_be_replaced,
497 access->grp_to_be_debug_replaced, access->grp_maybe_modified,
498 access->grp_not_necessarilly_dereferenced);
499 else
500 fprintf (f, ", write = %d, grp_total_scalarization = %d, "
501 "grp_partial_lhs = %d\n",
502 access->write, access->grp_total_scalarization,
503 access->grp_partial_lhs);
504 }
505
506 /* Dump a subtree rooted in ACCESS to file F, indent by LEVEL. */
507
508 static void
509 dump_access_tree_1 (FILE *f, struct access *access, int level)
510 {
511 do
512 {
513 int i;
514
515 for (i = 0; i < level; i++)
516 fputs ("* ", dump_file);
517
518 dump_access (f, access, true);
519
520 if (access->first_child)
521 dump_access_tree_1 (f, access->first_child, level + 1);
522
523 access = access->next_sibling;
524 }
525 while (access);
526 }
527
528 /* Dump all access trees for a variable, given the pointer to the first root in
529 ACCESS. */
530
531 static void
532 dump_access_tree (FILE *f, struct access *access)
533 {
534 for (; access; access = access->next_grp)
535 dump_access_tree_1 (f, access, 0);
536 }
537
538 /* Return true iff ACC is non-NULL and has subaccesses. */
539
540 static inline bool
541 access_has_children_p (struct access *acc)
542 {
543 return acc && acc->first_child;
544 }
545
546 /* Return true iff ACC is (partly) covered by at least one replacement. */
547
548 static bool
549 access_has_replacements_p (struct access *acc)
550 {
551 struct access *child;
552 if (acc->grp_to_be_replaced)
553 return true;
554 for (child = acc->first_child; child; child = child->next_sibling)
555 if (access_has_replacements_p (child))
556 return true;
557 return false;
558 }
559
560 /* Return a vector of pointers to accesses for the variable given in BASE or
561 NULL if there is none. */
562
563 static vec<access_p> *
564 get_base_access_vector (tree base)
565 {
566 return base_access_vec->get (base);
567 }
568
569 /* Find an access with required OFFSET and SIZE in a subtree of accesses rooted
570 in ACCESS. Return NULL if it cannot be found. */
571
572 static struct access *
573 find_access_in_subtree (struct access *access, HOST_WIDE_INT offset,
574 HOST_WIDE_INT size)
575 {
576 while (access && (access->offset != offset || access->size != size))
577 {
578 struct access *child = access->first_child;
579
580 while (child && (child->offset + child->size <= offset))
581 child = child->next_sibling;
582 access = child;
583 }
584
585 return access;
586 }
587
588 /* Return the first group representative for DECL or NULL if none exists. */
589
590 static struct access *
591 get_first_repr_for_decl (tree base)
592 {
593 vec<access_p> *access_vec;
594
595 access_vec = get_base_access_vector (base);
596 if (!access_vec)
597 return NULL;
598
599 return (*access_vec)[0];
600 }
601
602 /* Find an access representative for the variable BASE and given OFFSET and
603 SIZE. Requires that access trees have already been built. Return NULL if
604 it cannot be found. */
605
606 static struct access *
607 get_var_base_offset_size_access (tree base, HOST_WIDE_INT offset,
608 HOST_WIDE_INT size)
609 {
610 struct access *access;
611
612 access = get_first_repr_for_decl (base);
613 while (access && (access->offset + access->size <= offset))
614 access = access->next_grp;
615 if (!access)
616 return NULL;
617
618 return find_access_in_subtree (access, offset, size);
619 }
620
621 /* Add LINK to the linked list of assign links of RACC. */
622 static void
623 add_link_to_rhs (struct access *racc, struct assign_link *link)
624 {
625 gcc_assert (link->racc == racc);
626
627 if (!racc->first_link)
628 {
629 gcc_assert (!racc->last_link);
630 racc->first_link = link;
631 }
632 else
633 racc->last_link->next = link;
634
635 racc->last_link = link;
636 link->next = NULL;
637 }
638
639 /* Move all link structures in their linked list in OLD_RACC to the linked list
640 in NEW_RACC. */
641 static void
642 relink_to_new_repr (struct access *new_racc, struct access *old_racc)
643 {
644 if (!old_racc->first_link)
645 {
646 gcc_assert (!old_racc->last_link);
647 return;
648 }
649
650 if (new_racc->first_link)
651 {
652 gcc_assert (!new_racc->last_link->next);
653 gcc_assert (!old_racc->last_link || !old_racc->last_link->next);
654
655 new_racc->last_link->next = old_racc->first_link;
656 new_racc->last_link = old_racc->last_link;
657 }
658 else
659 {
660 gcc_assert (!new_racc->last_link);
661
662 new_racc->first_link = old_racc->first_link;
663 new_racc->last_link = old_racc->last_link;
664 }
665 old_racc->first_link = old_racc->last_link = NULL;
666 }
667
668 /* Add ACCESS to the work queue (which is actually a stack). */
669
670 static void
671 add_access_to_work_queue (struct access *access)
672 {
673 if (!access->grp_queued)
674 {
675 gcc_assert (!access->next_queued);
676 access->next_queued = work_queue_head;
677 access->grp_queued = 1;
678 work_queue_head = access;
679 }
680 }
681
682 /* Pop an access from the work queue, and return it, assuming there is one. */
683
684 static struct access *
685 pop_access_from_work_queue (void)
686 {
687 struct access *access = work_queue_head;
688
689 work_queue_head = access->next_queued;
690 access->next_queued = NULL;
691 access->grp_queued = 0;
692 return access;
693 }
694
695
696 /* Allocate necessary structures. */
697
698 static void
699 sra_initialize (void)
700 {
701 candidate_bitmap = BITMAP_ALLOC (NULL);
702 candidates = new hash_table<uid_decl_hasher>
703 (vec_safe_length (cfun->local_decls) / 2);
704 should_scalarize_away_bitmap = BITMAP_ALLOC (NULL);
705 cannot_scalarize_away_bitmap = BITMAP_ALLOC (NULL);
706 gcc_obstack_init (&name_obstack);
707 base_access_vec = new hash_map<tree, auto_vec<access_p> >;
708 memset (&sra_stats, 0, sizeof (sra_stats));
709 encountered_apply_args = false;
710 encountered_recursive_call = false;
711 encountered_unchangable_recursive_call = false;
712 }
713
714 /* Deallocate all general structures. */
715
716 static void
717 sra_deinitialize (void)
718 {
719 BITMAP_FREE (candidate_bitmap);
720 delete candidates;
721 candidates = NULL;
722 BITMAP_FREE (should_scalarize_away_bitmap);
723 BITMAP_FREE (cannot_scalarize_away_bitmap);
724 access::pool.release ();
725 assign_link::pool.release ();
726 obstack_free (&name_obstack, NULL);
727
728 delete base_access_vec;
729 }
730
731 /* Remove DECL from candidates for SRA and write REASON to the dump file if
732 there is one. */
733 static void
734 disqualify_candidate (tree decl, const char *reason)
735 {
736 if (bitmap_clear_bit (candidate_bitmap, DECL_UID (decl)))
737 candidates->remove_elt_with_hash (decl, DECL_UID (decl));
738
739 if (dump_file && (dump_flags & TDF_DETAILS))
740 {
741 fprintf (dump_file, "! Disqualifying ");
742 print_generic_expr (dump_file, decl, 0);
743 fprintf (dump_file, " - %s\n", reason);
744 }
745 }
746
747 /* Return true iff the type contains a field or an element which does not allow
748 scalarization. */
749
750 static bool
751 type_internals_preclude_sra_p (tree type, const char **msg)
752 {
753 tree fld;
754 tree et;
755
756 switch (TREE_CODE (type))
757 {
758 case RECORD_TYPE:
759 case UNION_TYPE:
760 case QUAL_UNION_TYPE:
761 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
762 if (TREE_CODE (fld) == FIELD_DECL)
763 {
764 tree ft = TREE_TYPE (fld);
765
766 if (TREE_THIS_VOLATILE (fld))
767 {
768 *msg = "volatile structure field";
769 return true;
770 }
771 if (!DECL_FIELD_OFFSET (fld))
772 {
773 *msg = "no structure field offset";
774 return true;
775 }
776 if (!DECL_SIZE (fld))
777 {
778 *msg = "zero structure field size";
779 return true;
780 }
781 if (!tree_fits_uhwi_p (DECL_FIELD_OFFSET (fld)))
782 {
783 *msg = "structure field offset not fixed";
784 return true;
785 }
786 if (!tree_fits_uhwi_p (DECL_SIZE (fld)))
787 {
788 *msg = "structure field size not fixed";
789 return true;
790 }
791 if (!tree_fits_shwi_p (bit_position (fld)))
792 {
793 *msg = "structure field size too big";
794 return true;
795 }
796 if (AGGREGATE_TYPE_P (ft)
797 && int_bit_position (fld) % BITS_PER_UNIT != 0)
798 {
799 *msg = "structure field is bit field";
800 return true;
801 }
802
803 if (AGGREGATE_TYPE_P (ft) && type_internals_preclude_sra_p (ft, msg))
804 return true;
805 }
806
807 return false;
808
809 case ARRAY_TYPE:
810 et = TREE_TYPE (type);
811
812 if (TYPE_VOLATILE (et))
813 {
814 *msg = "element type is volatile";
815 return true;
816 }
817
818 if (AGGREGATE_TYPE_P (et) && type_internals_preclude_sra_p (et, msg))
819 return true;
820
821 return false;
822
823 default:
824 return false;
825 }
826 }
827
828 /* If T is an SSA_NAME, return NULL if it is not a default def or return its
829 base variable if it is. Return T if it is not an SSA_NAME. */
830
831 static tree
832 get_ssa_base_param (tree t)
833 {
834 if (TREE_CODE (t) == SSA_NAME)
835 {
836 if (SSA_NAME_IS_DEFAULT_DEF (t))
837 return SSA_NAME_VAR (t);
838 else
839 return NULL_TREE;
840 }
841 return t;
842 }
843
844 /* Mark a dereference of BASE of distance DIST in a basic block tht STMT
845 belongs to, unless the BB has already been marked as a potentially
846 final. */
847
848 static void
849 mark_parm_dereference (tree base, HOST_WIDE_INT dist, gimple stmt)
850 {
851 basic_block bb = gimple_bb (stmt);
852 int idx, parm_index = 0;
853 tree parm;
854
855 if (bitmap_bit_p (final_bbs, bb->index))
856 return;
857
858 for (parm = DECL_ARGUMENTS (current_function_decl);
859 parm && parm != base;
860 parm = DECL_CHAIN (parm))
861 parm_index++;
862
863 gcc_assert (parm_index < func_param_count);
864
865 idx = bb->index * func_param_count + parm_index;
866 if (bb_dereferences[idx] < dist)
867 bb_dereferences[idx] = dist;
868 }
869
870 /* Allocate an access structure for BASE, OFFSET and SIZE, clear it, fill in
871 the three fields. Also add it to the vector of accesses corresponding to
872 the base. Finally, return the new access. */
873
874 static struct access *
875 create_access_1 (tree base, HOST_WIDE_INT offset, HOST_WIDE_INT size)
876 {
877 struct access *access = new struct access ();
878
879 memset (access, 0, sizeof (struct access));
880 access->base = base;
881 access->offset = offset;
882 access->size = size;
883
884 base_access_vec->get_or_insert (base).safe_push (access);
885
886 return access;
887 }
888
889 /* Create and insert access for EXPR. Return created access, or NULL if it is
890 not possible. */
891
892 static struct access *
893 create_access (tree expr, gimple stmt, bool write)
894 {
895 struct access *access;
896 HOST_WIDE_INT offset, size, max_size;
897 tree base = expr;
898 bool ptr, unscalarizable_region = false;
899
900 base = get_ref_base_and_extent (expr, &offset, &size, &max_size);
901
902 if (sra_mode == SRA_MODE_EARLY_IPA
903 && TREE_CODE (base) == MEM_REF)
904 {
905 base = get_ssa_base_param (TREE_OPERAND (base, 0));
906 if (!base)
907 return NULL;
908 ptr = true;
909 }
910 else
911 ptr = false;
912
913 if (!DECL_P (base) || !bitmap_bit_p (candidate_bitmap, DECL_UID (base)))
914 return NULL;
915
916 if (sra_mode == SRA_MODE_EARLY_IPA)
917 {
918 if (size < 0 || size != max_size)
919 {
920 disqualify_candidate (base, "Encountered a variable sized access.");
921 return NULL;
922 }
923 if (TREE_CODE (expr) == COMPONENT_REF
924 && DECL_BIT_FIELD (TREE_OPERAND (expr, 1)))
925 {
926 disqualify_candidate (base, "Encountered a bit-field access.");
927 return NULL;
928 }
929 gcc_checking_assert ((offset % BITS_PER_UNIT) == 0);
930
931 if (ptr)
932 mark_parm_dereference (base, offset + size, stmt);
933 }
934 else
935 {
936 if (size != max_size)
937 {
938 size = max_size;
939 unscalarizable_region = true;
940 }
941 if (size < 0)
942 {
943 disqualify_candidate (base, "Encountered an unconstrained access.");
944 return NULL;
945 }
946 }
947
948 access = create_access_1 (base, offset, size);
949 access->expr = expr;
950 access->type = TREE_TYPE (expr);
951 access->write = write;
952 access->grp_unscalarizable_region = unscalarizable_region;
953 access->stmt = stmt;
954
955 if (TREE_CODE (expr) == COMPONENT_REF
956 && DECL_NONADDRESSABLE_P (TREE_OPERAND (expr, 1)))
957 access->non_addressable = 1;
958
959 return access;
960 }
961
962
963 /* Return true iff TYPE is a RECORD_TYPE with fields that are either of gimple
964 register types or (recursively) records with only these two kinds of fields.
965 It also returns false if any of these records contains a bit-field. */
966
967 static bool
968 type_consists_of_records_p (tree type)
969 {
970 tree fld;
971
972 if (TREE_CODE (type) != RECORD_TYPE)
973 return false;
974
975 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
976 if (TREE_CODE (fld) == FIELD_DECL)
977 {
978 tree ft = TREE_TYPE (fld);
979
980 if (DECL_BIT_FIELD (fld))
981 return false;
982
983 if (!is_gimple_reg_type (ft)
984 && !type_consists_of_records_p (ft))
985 return false;
986 }
987
988 return true;
989 }
990
991 /* Create total_scalarization accesses for all scalar type fields in DECL that
992 must be of a RECORD_TYPE conforming to type_consists_of_records_p. BASE
993 must be the top-most VAR_DECL representing the variable, OFFSET must be the
994 offset of DECL within BASE. REF must be the memory reference expression for
995 the given decl. */
996
997 static void
998 completely_scalarize_record (tree base, tree decl, HOST_WIDE_INT offset,
999 tree ref)
1000 {
1001 tree fld, decl_type = TREE_TYPE (decl);
1002
1003 for (fld = TYPE_FIELDS (decl_type); fld; fld = DECL_CHAIN (fld))
1004 if (TREE_CODE (fld) == FIELD_DECL)
1005 {
1006 HOST_WIDE_INT pos = offset + int_bit_position (fld);
1007 tree ft = TREE_TYPE (fld);
1008 tree nref = build3 (COMPONENT_REF, TREE_TYPE (fld), ref, fld,
1009 NULL_TREE);
1010
1011 if (is_gimple_reg_type (ft))
1012 {
1013 struct access *access;
1014 HOST_WIDE_INT size;
1015
1016 size = tree_to_uhwi (DECL_SIZE (fld));
1017 access = create_access_1 (base, pos, size);
1018 access->expr = nref;
1019 access->type = ft;
1020 access->grp_total_scalarization = 1;
1021 /* Accesses for intraprocedural SRA can have their stmt NULL. */
1022 }
1023 else
1024 completely_scalarize_record (base, fld, pos, nref);
1025 }
1026 }
1027
1028 /* Create total_scalarization accesses for all scalar type fields in VAR and
1029 for VAR a a whole. VAR must be of a RECORD_TYPE conforming to
1030 type_consists_of_records_p. */
1031
1032 static void
1033 completely_scalarize_var (tree var)
1034 {
1035 HOST_WIDE_INT size = tree_to_uhwi (DECL_SIZE (var));
1036 struct access *access;
1037
1038 access = create_access_1 (var, 0, size);
1039 access->expr = var;
1040 access->type = TREE_TYPE (var);
1041 access->grp_total_scalarization = 1;
1042
1043 completely_scalarize_record (var, var, 0, var);
1044 }
1045
1046 /* Return true if REF has an VIEW_CONVERT_EXPR somewhere in it. */
1047
1048 static inline bool
1049 contains_view_convert_expr_p (const_tree ref)
1050 {
1051 while (handled_component_p (ref))
1052 {
1053 if (TREE_CODE (ref) == VIEW_CONVERT_EXPR)
1054 return true;
1055 ref = TREE_OPERAND (ref, 0);
1056 }
1057
1058 return false;
1059 }
1060
1061 /* Search the given tree for a declaration by skipping handled components and
1062 exclude it from the candidates. */
1063
1064 static void
1065 disqualify_base_of_expr (tree t, const char *reason)
1066 {
1067 t = get_base_address (t);
1068 if (sra_mode == SRA_MODE_EARLY_IPA
1069 && TREE_CODE (t) == MEM_REF)
1070 t = get_ssa_base_param (TREE_OPERAND (t, 0));
1071
1072 if (t && DECL_P (t))
1073 disqualify_candidate (t, reason);
1074 }
1075
1076 /* Scan expression EXPR and create access structures for all accesses to
1077 candidates for scalarization. Return the created access or NULL if none is
1078 created. */
1079
1080 static struct access *
1081 build_access_from_expr_1 (tree expr, gimple stmt, bool write)
1082 {
1083 struct access *ret = NULL;
1084 bool partial_ref;
1085
1086 if (TREE_CODE (expr) == BIT_FIELD_REF
1087 || TREE_CODE (expr) == IMAGPART_EXPR
1088 || TREE_CODE (expr) == REALPART_EXPR)
1089 {
1090 expr = TREE_OPERAND (expr, 0);
1091 partial_ref = true;
1092 }
1093 else
1094 partial_ref = false;
1095
1096 /* We need to dive through V_C_Es in order to get the size of its parameter
1097 and not the result type. Ada produces such statements. We are also
1098 capable of handling the topmost V_C_E but not any of those buried in other
1099 handled components. */
1100 if (TREE_CODE (expr) == VIEW_CONVERT_EXPR)
1101 expr = TREE_OPERAND (expr, 0);
1102
1103 if (contains_view_convert_expr_p (expr))
1104 {
1105 disqualify_base_of_expr (expr, "V_C_E under a different handled "
1106 "component.");
1107 return NULL;
1108 }
1109 if (TREE_THIS_VOLATILE (expr))
1110 {
1111 disqualify_base_of_expr (expr, "part of a volatile reference.");
1112 return NULL;
1113 }
1114
1115 switch (TREE_CODE (expr))
1116 {
1117 case MEM_REF:
1118 if (TREE_CODE (TREE_OPERAND (expr, 0)) != ADDR_EXPR
1119 && sra_mode != SRA_MODE_EARLY_IPA)
1120 return NULL;
1121 /* fall through */
1122 case VAR_DECL:
1123 case PARM_DECL:
1124 case RESULT_DECL:
1125 case COMPONENT_REF:
1126 case ARRAY_REF:
1127 case ARRAY_RANGE_REF:
1128 ret = create_access (expr, stmt, write);
1129 break;
1130
1131 default:
1132 break;
1133 }
1134
1135 if (write && partial_ref && ret)
1136 ret->grp_partial_lhs = 1;
1137
1138 return ret;
1139 }
1140
1141 /* Scan expression EXPR and create access structures for all accesses to
1142 candidates for scalarization. Return true if any access has been inserted.
1143 STMT must be the statement from which the expression is taken, WRITE must be
1144 true if the expression is a store and false otherwise. */
1145
1146 static bool
1147 build_access_from_expr (tree expr, gimple stmt, bool write)
1148 {
1149 struct access *access;
1150
1151 access = build_access_from_expr_1 (expr, stmt, write);
1152 if (access)
1153 {
1154 /* This means the aggregate is accesses as a whole in a way other than an
1155 assign statement and thus cannot be removed even if we had a scalar
1156 replacement for everything. */
1157 if (cannot_scalarize_away_bitmap)
1158 bitmap_set_bit (cannot_scalarize_away_bitmap, DECL_UID (access->base));
1159 return true;
1160 }
1161 return false;
1162 }
1163
1164 /* Return the single non-EH successor edge of BB or NULL if there is none or
1165 more than one. */
1166
1167 static edge
1168 single_non_eh_succ (basic_block bb)
1169 {
1170 edge e, res = NULL;
1171 edge_iterator ei;
1172
1173 FOR_EACH_EDGE (e, ei, bb->succs)
1174 if (!(e->flags & EDGE_EH))
1175 {
1176 if (res)
1177 return NULL;
1178 res = e;
1179 }
1180
1181 return res;
1182 }
1183
1184 /* Disqualify LHS and RHS for scalarization if STMT has to terminate its BB and
1185 there is no alternative spot where to put statements SRA might need to
1186 generate after it. The spot we are looking for is an edge leading to a
1187 single non-EH successor, if it exists and is indeed single. RHS may be
1188 NULL, in that case ignore it. */
1189
1190 static bool
1191 disqualify_if_bad_bb_terminating_stmt (gimple stmt, tree lhs, tree rhs)
1192 {
1193 if ((sra_mode == SRA_MODE_EARLY_INTRA || sra_mode == SRA_MODE_INTRA)
1194 && stmt_ends_bb_p (stmt))
1195 {
1196 if (single_non_eh_succ (gimple_bb (stmt)))
1197 return false;
1198
1199 disqualify_base_of_expr (lhs, "LHS of a throwing stmt.");
1200 if (rhs)
1201 disqualify_base_of_expr (rhs, "RHS of a throwing stmt.");
1202 return true;
1203 }
1204 return false;
1205 }
1206
1207 /* Scan expressions occurring in STMT, create access structures for all accesses
1208 to candidates for scalarization and remove those candidates which occur in
1209 statements or expressions that prevent them from being split apart. Return
1210 true if any access has been inserted. */
1211
1212 static bool
1213 build_accesses_from_assign (gimple stmt)
1214 {
1215 tree lhs, rhs;
1216 struct access *lacc, *racc;
1217
1218 if (!gimple_assign_single_p (stmt)
1219 /* Scope clobbers don't influence scalarization. */
1220 || gimple_clobber_p (stmt))
1221 return false;
1222
1223 lhs = gimple_assign_lhs (stmt);
1224 rhs = gimple_assign_rhs1 (stmt);
1225
1226 if (disqualify_if_bad_bb_terminating_stmt (stmt, lhs, rhs))
1227 return false;
1228
1229 racc = build_access_from_expr_1 (rhs, stmt, false);
1230 lacc = build_access_from_expr_1 (lhs, stmt, true);
1231
1232 if (lacc)
1233 lacc->grp_assignment_write = 1;
1234
1235 if (racc)
1236 {
1237 racc->grp_assignment_read = 1;
1238 if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
1239 && !is_gimple_reg_type (racc->type))
1240 bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
1241 }
1242
1243 if (lacc && racc
1244 && (sra_mode == SRA_MODE_EARLY_INTRA || sra_mode == SRA_MODE_INTRA)
1245 && !lacc->grp_unscalarizable_region
1246 && !racc->grp_unscalarizable_region
1247 && AGGREGATE_TYPE_P (TREE_TYPE (lhs))
1248 && lacc->size == racc->size
1249 && useless_type_conversion_p (lacc->type, racc->type))
1250 {
1251 struct assign_link *link;
1252
1253 link = new assign_link;
1254 memset (link, 0, sizeof (struct assign_link));
1255
1256 link->lacc = lacc;
1257 link->racc = racc;
1258
1259 add_link_to_rhs (racc, link);
1260 }
1261
1262 return lacc || racc;
1263 }
1264
1265 /* Callback of walk_stmt_load_store_addr_ops visit_addr used to determine
1266 GIMPLE_ASM operands with memory constrains which cannot be scalarized. */
1267
1268 static bool
1269 asm_visit_addr (gimple, tree op, tree, void *)
1270 {
1271 op = get_base_address (op);
1272 if (op
1273 && DECL_P (op))
1274 disqualify_candidate (op, "Non-scalarizable GIMPLE_ASM operand.");
1275
1276 return false;
1277 }
1278
1279 /* Return true iff callsite CALL has at least as many actual arguments as there
1280 are formal parameters of the function currently processed by IPA-SRA and
1281 that their types match. */
1282
1283 static inline bool
1284 callsite_arguments_match_p (gimple call)
1285 {
1286 if (gimple_call_num_args (call) < (unsigned) func_param_count)
1287 return false;
1288
1289 tree parm;
1290 int i;
1291 for (parm = DECL_ARGUMENTS (current_function_decl), i = 0;
1292 parm;
1293 parm = DECL_CHAIN (parm), i++)
1294 {
1295 tree arg = gimple_call_arg (call, i);
1296 if (!useless_type_conversion_p (TREE_TYPE (parm), TREE_TYPE (arg)))
1297 return false;
1298 }
1299 return true;
1300 }
1301
1302 /* Scan function and look for interesting expressions and create access
1303 structures for them. Return true iff any access is created. */
1304
1305 static bool
1306 scan_function (void)
1307 {
1308 basic_block bb;
1309 bool ret = false;
1310
1311 FOR_EACH_BB_FN (bb, cfun)
1312 {
1313 gimple_stmt_iterator gsi;
1314 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1315 {
1316 gimple stmt = gsi_stmt (gsi);
1317 tree t;
1318 unsigned i;
1319
1320 if (final_bbs && stmt_can_throw_external (stmt))
1321 bitmap_set_bit (final_bbs, bb->index);
1322 switch (gimple_code (stmt))
1323 {
1324 case GIMPLE_RETURN:
1325 t = gimple_return_retval (as_a <greturn *> (stmt));
1326 if (t != NULL_TREE)
1327 ret |= build_access_from_expr (t, stmt, false);
1328 if (final_bbs)
1329 bitmap_set_bit (final_bbs, bb->index);
1330 break;
1331
1332 case GIMPLE_ASSIGN:
1333 ret |= build_accesses_from_assign (stmt);
1334 break;
1335
1336 case GIMPLE_CALL:
1337 for (i = 0; i < gimple_call_num_args (stmt); i++)
1338 ret |= build_access_from_expr (gimple_call_arg (stmt, i),
1339 stmt, false);
1340
1341 if (sra_mode == SRA_MODE_EARLY_IPA)
1342 {
1343 tree dest = gimple_call_fndecl (stmt);
1344 int flags = gimple_call_flags (stmt);
1345
1346 if (dest)
1347 {
1348 if (DECL_BUILT_IN_CLASS (dest) == BUILT_IN_NORMAL
1349 && DECL_FUNCTION_CODE (dest) == BUILT_IN_APPLY_ARGS)
1350 encountered_apply_args = true;
1351 if (recursive_call_p (current_function_decl, dest))
1352 {
1353 encountered_recursive_call = true;
1354 if (!callsite_arguments_match_p (stmt))
1355 encountered_unchangable_recursive_call = true;
1356 }
1357 }
1358
1359 if (final_bbs
1360 && (flags & (ECF_CONST | ECF_PURE)) == 0)
1361 bitmap_set_bit (final_bbs, bb->index);
1362 }
1363
1364 t = gimple_call_lhs (stmt);
1365 if (t && !disqualify_if_bad_bb_terminating_stmt (stmt, t, NULL))
1366 ret |= build_access_from_expr (t, stmt, true);
1367 break;
1368
1369 case GIMPLE_ASM:
1370 {
1371 gasm *asm_stmt = as_a <gasm *> (stmt);
1372 walk_stmt_load_store_addr_ops (asm_stmt, NULL, NULL, NULL,
1373 asm_visit_addr);
1374 if (final_bbs)
1375 bitmap_set_bit (final_bbs, bb->index);
1376
1377 for (i = 0; i < gimple_asm_ninputs (asm_stmt); i++)
1378 {
1379 t = TREE_VALUE (gimple_asm_input_op (asm_stmt, i));
1380 ret |= build_access_from_expr (t, asm_stmt, false);
1381 }
1382 for (i = 0; i < gimple_asm_noutputs (asm_stmt); i++)
1383 {
1384 t = TREE_VALUE (gimple_asm_output_op (asm_stmt, i));
1385 ret |= build_access_from_expr (t, asm_stmt, true);
1386 }
1387 }
1388 break;
1389
1390 default:
1391 break;
1392 }
1393 }
1394 }
1395
1396 return ret;
1397 }
1398
1399 /* Helper of QSORT function. There are pointers to accesses in the array. An
1400 access is considered smaller than another if it has smaller offset or if the
1401 offsets are the same but is size is bigger. */
1402
1403 static int
1404 compare_access_positions (const void *a, const void *b)
1405 {
1406 const access_p *fp1 = (const access_p *) a;
1407 const access_p *fp2 = (const access_p *) b;
1408 const access_p f1 = *fp1;
1409 const access_p f2 = *fp2;
1410
1411 if (f1->offset != f2->offset)
1412 return f1->offset < f2->offset ? -1 : 1;
1413
1414 if (f1->size == f2->size)
1415 {
1416 if (f1->type == f2->type)
1417 return 0;
1418 /* Put any non-aggregate type before any aggregate type. */
1419 else if (!is_gimple_reg_type (f1->type)
1420 && is_gimple_reg_type (f2->type))
1421 return 1;
1422 else if (is_gimple_reg_type (f1->type)
1423 && !is_gimple_reg_type (f2->type))
1424 return -1;
1425 /* Put any complex or vector type before any other scalar type. */
1426 else if (TREE_CODE (f1->type) != COMPLEX_TYPE
1427 && TREE_CODE (f1->type) != VECTOR_TYPE
1428 && (TREE_CODE (f2->type) == COMPLEX_TYPE
1429 || TREE_CODE (f2->type) == VECTOR_TYPE))
1430 return 1;
1431 else if ((TREE_CODE (f1->type) == COMPLEX_TYPE
1432 || TREE_CODE (f1->type) == VECTOR_TYPE)
1433 && TREE_CODE (f2->type) != COMPLEX_TYPE
1434 && TREE_CODE (f2->type) != VECTOR_TYPE)
1435 return -1;
1436 /* Put the integral type with the bigger precision first. */
1437 else if (INTEGRAL_TYPE_P (f1->type)
1438 && INTEGRAL_TYPE_P (f2->type))
1439 return TYPE_PRECISION (f2->type) - TYPE_PRECISION (f1->type);
1440 /* Put any integral type with non-full precision last. */
1441 else if (INTEGRAL_TYPE_P (f1->type)
1442 && (TREE_INT_CST_LOW (TYPE_SIZE (f1->type))
1443 != TYPE_PRECISION (f1->type)))
1444 return 1;
1445 else if (INTEGRAL_TYPE_P (f2->type)
1446 && (TREE_INT_CST_LOW (TYPE_SIZE (f2->type))
1447 != TYPE_PRECISION (f2->type)))
1448 return -1;
1449 /* Stabilize the sort. */
1450 return TYPE_UID (f1->type) - TYPE_UID (f2->type);
1451 }
1452
1453 /* We want the bigger accesses first, thus the opposite operator in the next
1454 line: */
1455 return f1->size > f2->size ? -1 : 1;
1456 }
1457
1458
1459 /* Append a name of the declaration to the name obstack. A helper function for
1460 make_fancy_name. */
1461
1462 static void
1463 make_fancy_decl_name (tree decl)
1464 {
1465 char buffer[32];
1466
1467 tree name = DECL_NAME (decl);
1468 if (name)
1469 obstack_grow (&name_obstack, IDENTIFIER_POINTER (name),
1470 IDENTIFIER_LENGTH (name));
1471 else
1472 {
1473 sprintf (buffer, "D%u", DECL_UID (decl));
1474 obstack_grow (&name_obstack, buffer, strlen (buffer));
1475 }
1476 }
1477
1478 /* Helper for make_fancy_name. */
1479
1480 static void
1481 make_fancy_name_1 (tree expr)
1482 {
1483 char buffer[32];
1484 tree index;
1485
1486 if (DECL_P (expr))
1487 {
1488 make_fancy_decl_name (expr);
1489 return;
1490 }
1491
1492 switch (TREE_CODE (expr))
1493 {
1494 case COMPONENT_REF:
1495 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1496 obstack_1grow (&name_obstack, '$');
1497 make_fancy_decl_name (TREE_OPERAND (expr, 1));
1498 break;
1499
1500 case ARRAY_REF:
1501 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1502 obstack_1grow (&name_obstack, '$');
1503 /* Arrays with only one element may not have a constant as their
1504 index. */
1505 index = TREE_OPERAND (expr, 1);
1506 if (TREE_CODE (index) != INTEGER_CST)
1507 break;
1508 sprintf (buffer, HOST_WIDE_INT_PRINT_DEC, TREE_INT_CST_LOW (index));
1509 obstack_grow (&name_obstack, buffer, strlen (buffer));
1510 break;
1511
1512 case ADDR_EXPR:
1513 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1514 break;
1515
1516 case MEM_REF:
1517 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1518 if (!integer_zerop (TREE_OPERAND (expr, 1)))
1519 {
1520 obstack_1grow (&name_obstack, '$');
1521 sprintf (buffer, HOST_WIDE_INT_PRINT_DEC,
1522 TREE_INT_CST_LOW (TREE_OPERAND (expr, 1)));
1523 obstack_grow (&name_obstack, buffer, strlen (buffer));
1524 }
1525 break;
1526
1527 case BIT_FIELD_REF:
1528 case REALPART_EXPR:
1529 case IMAGPART_EXPR:
1530 gcc_unreachable (); /* we treat these as scalars. */
1531 break;
1532 default:
1533 break;
1534 }
1535 }
1536
1537 /* Create a human readable name for replacement variable of ACCESS. */
1538
1539 static char *
1540 make_fancy_name (tree expr)
1541 {
1542 make_fancy_name_1 (expr);
1543 obstack_1grow (&name_obstack, '\0');
1544 return XOBFINISH (&name_obstack, char *);
1545 }
1546
1547 /* Construct a MEM_REF that would reference a part of aggregate BASE of type
1548 EXP_TYPE at the given OFFSET. If BASE is something for which
1549 get_addr_base_and_unit_offset returns NULL, gsi must be non-NULL and is used
1550 to insert new statements either before or below the current one as specified
1551 by INSERT_AFTER. This function is not capable of handling bitfields.
1552
1553 BASE must be either a declaration or a memory reference that has correct
1554 alignment ifformation embeded in it (e.g. a pre-existing one in SRA). */
1555
1556 tree
1557 build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
1558 tree exp_type, gimple_stmt_iterator *gsi,
1559 bool insert_after)
1560 {
1561 tree prev_base = base;
1562 tree off;
1563 tree mem_ref;
1564 HOST_WIDE_INT base_offset;
1565 unsigned HOST_WIDE_INT misalign;
1566 unsigned int align;
1567
1568 gcc_checking_assert (offset % BITS_PER_UNIT == 0);
1569 get_object_alignment_1 (base, &align, &misalign);
1570 base = get_addr_base_and_unit_offset (base, &base_offset);
1571
1572 /* get_addr_base_and_unit_offset returns NULL for references with a variable
1573 offset such as array[var_index]. */
1574 if (!base)
1575 {
1576 gassign *stmt;
1577 tree tmp, addr;
1578
1579 gcc_checking_assert (gsi);
1580 tmp = make_ssa_name (build_pointer_type (TREE_TYPE (prev_base)));
1581 addr = build_fold_addr_expr (unshare_expr (prev_base));
1582 STRIP_USELESS_TYPE_CONVERSION (addr);
1583 stmt = gimple_build_assign (tmp, addr);
1584 gimple_set_location (stmt, loc);
1585 if (insert_after)
1586 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
1587 else
1588 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
1589
1590 off = build_int_cst (reference_alias_ptr_type (prev_base),
1591 offset / BITS_PER_UNIT);
1592 base = tmp;
1593 }
1594 else if (TREE_CODE (base) == MEM_REF)
1595 {
1596 off = build_int_cst (TREE_TYPE (TREE_OPERAND (base, 1)),
1597 base_offset + offset / BITS_PER_UNIT);
1598 off = int_const_binop (PLUS_EXPR, TREE_OPERAND (base, 1), off);
1599 base = unshare_expr (TREE_OPERAND (base, 0));
1600 }
1601 else
1602 {
1603 off = build_int_cst (reference_alias_ptr_type (base),
1604 base_offset + offset / BITS_PER_UNIT);
1605 base = build_fold_addr_expr (unshare_expr (base));
1606 }
1607
1608 misalign = (misalign + offset) & (align - 1);
1609 if (misalign != 0)
1610 align = (misalign & -misalign);
1611 if (align != TYPE_ALIGN (exp_type))
1612 exp_type = build_aligned_type (exp_type, align);
1613
1614 mem_ref = fold_build2_loc (loc, MEM_REF, exp_type, base, off);
1615 if (TREE_THIS_VOLATILE (prev_base))
1616 TREE_THIS_VOLATILE (mem_ref) = 1;
1617 if (TREE_SIDE_EFFECTS (prev_base))
1618 TREE_SIDE_EFFECTS (mem_ref) = 1;
1619 return mem_ref;
1620 }
1621
1622 /* Construct a memory reference to a part of an aggregate BASE at the given
1623 OFFSET and of the same type as MODEL. In case this is a reference to a
1624 bit-field, the function will replicate the last component_ref of model's
1625 expr to access it. GSI and INSERT_AFTER have the same meaning as in
1626 build_ref_for_offset. */
1627
1628 static tree
1629 build_ref_for_model (location_t loc, tree base, HOST_WIDE_INT offset,
1630 struct access *model, gimple_stmt_iterator *gsi,
1631 bool insert_after)
1632 {
1633 if (TREE_CODE (model->expr) == COMPONENT_REF
1634 && DECL_BIT_FIELD (TREE_OPERAND (model->expr, 1)))
1635 {
1636 /* This access represents a bit-field. */
1637 tree t, exp_type, fld = TREE_OPERAND (model->expr, 1);
1638
1639 offset -= int_bit_position (fld);
1640 exp_type = TREE_TYPE (TREE_OPERAND (model->expr, 0));
1641 t = build_ref_for_offset (loc, base, offset, exp_type, gsi, insert_after);
1642 return fold_build3_loc (loc, COMPONENT_REF, TREE_TYPE (fld), t, fld,
1643 NULL_TREE);
1644 }
1645 else
1646 return build_ref_for_offset (loc, base, offset, model->type,
1647 gsi, insert_after);
1648 }
1649
1650 /* Attempt to build a memory reference that we could but into a gimple
1651 debug_bind statement. Similar to build_ref_for_model but punts if it has to
1652 create statements and return s NULL instead. This function also ignores
1653 alignment issues and so its results should never end up in non-debug
1654 statements. */
1655
1656 static tree
1657 build_debug_ref_for_model (location_t loc, tree base, HOST_WIDE_INT offset,
1658 struct access *model)
1659 {
1660 HOST_WIDE_INT base_offset;
1661 tree off;
1662
1663 if (TREE_CODE (model->expr) == COMPONENT_REF
1664 && DECL_BIT_FIELD (TREE_OPERAND (model->expr, 1)))
1665 return NULL_TREE;
1666
1667 base = get_addr_base_and_unit_offset (base, &base_offset);
1668 if (!base)
1669 return NULL_TREE;
1670 if (TREE_CODE (base) == MEM_REF)
1671 {
1672 off = build_int_cst (TREE_TYPE (TREE_OPERAND (base, 1)),
1673 base_offset + offset / BITS_PER_UNIT);
1674 off = int_const_binop (PLUS_EXPR, TREE_OPERAND (base, 1), off);
1675 base = unshare_expr (TREE_OPERAND (base, 0));
1676 }
1677 else
1678 {
1679 off = build_int_cst (reference_alias_ptr_type (base),
1680 base_offset + offset / BITS_PER_UNIT);
1681 base = build_fold_addr_expr (unshare_expr (base));
1682 }
1683
1684 return fold_build2_loc (loc, MEM_REF, model->type, base, off);
1685 }
1686
1687 /* Construct a memory reference consisting of component_refs and array_refs to
1688 a part of an aggregate *RES (which is of type TYPE). The requested part
1689 should have type EXP_TYPE at be the given OFFSET. This function might not
1690 succeed, it returns true when it does and only then *RES points to something
1691 meaningful. This function should be used only to build expressions that we
1692 might need to present to user (e.g. in warnings). In all other situations,
1693 build_ref_for_model or build_ref_for_offset should be used instead. */
1694
1695 static bool
1696 build_user_friendly_ref_for_offset (tree *res, tree type, HOST_WIDE_INT offset,
1697 tree exp_type)
1698 {
1699 while (1)
1700 {
1701 tree fld;
1702 tree tr_size, index, minidx;
1703 HOST_WIDE_INT el_size;
1704
1705 if (offset == 0 && exp_type
1706 && types_compatible_p (exp_type, type))
1707 return true;
1708
1709 switch (TREE_CODE (type))
1710 {
1711 case UNION_TYPE:
1712 case QUAL_UNION_TYPE:
1713 case RECORD_TYPE:
1714 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
1715 {
1716 HOST_WIDE_INT pos, size;
1717 tree tr_pos, expr, *expr_ptr;
1718
1719 if (TREE_CODE (fld) != FIELD_DECL)
1720 continue;
1721
1722 tr_pos = bit_position (fld);
1723 if (!tr_pos || !tree_fits_uhwi_p (tr_pos))
1724 continue;
1725 pos = tree_to_uhwi (tr_pos);
1726 gcc_assert (TREE_CODE (type) == RECORD_TYPE || pos == 0);
1727 tr_size = DECL_SIZE (fld);
1728 if (!tr_size || !tree_fits_uhwi_p (tr_size))
1729 continue;
1730 size = tree_to_uhwi (tr_size);
1731 if (size == 0)
1732 {
1733 if (pos != offset)
1734 continue;
1735 }
1736 else if (pos > offset || (pos + size) <= offset)
1737 continue;
1738
1739 expr = build3 (COMPONENT_REF, TREE_TYPE (fld), *res, fld,
1740 NULL_TREE);
1741 expr_ptr = &expr;
1742 if (build_user_friendly_ref_for_offset (expr_ptr, TREE_TYPE (fld),
1743 offset - pos, exp_type))
1744 {
1745 *res = expr;
1746 return true;
1747 }
1748 }
1749 return false;
1750
1751 case ARRAY_TYPE:
1752 tr_size = TYPE_SIZE (TREE_TYPE (type));
1753 if (!tr_size || !tree_fits_uhwi_p (tr_size))
1754 return false;
1755 el_size = tree_to_uhwi (tr_size);
1756
1757 minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (type));
1758 if (TREE_CODE (minidx) != INTEGER_CST || el_size == 0)
1759 return false;
1760 index = build_int_cst (TYPE_DOMAIN (type), offset / el_size);
1761 if (!integer_zerop (minidx))
1762 index = int_const_binop (PLUS_EXPR, index, minidx);
1763 *res = build4 (ARRAY_REF, TREE_TYPE (type), *res, index,
1764 NULL_TREE, NULL_TREE);
1765 offset = offset % el_size;
1766 type = TREE_TYPE (type);
1767 break;
1768
1769 default:
1770 if (offset != 0)
1771 return false;
1772
1773 if (exp_type)
1774 return false;
1775 else
1776 return true;
1777 }
1778 }
1779 }
1780
1781 /* Return true iff TYPE is stdarg va_list type. */
1782
1783 static inline bool
1784 is_va_list_type (tree type)
1785 {
1786 return TYPE_MAIN_VARIANT (type) == TYPE_MAIN_VARIANT (va_list_type_node);
1787 }
1788
1789 /* Print message to dump file why a variable was rejected. */
1790
1791 static void
1792 reject (tree var, const char *msg)
1793 {
1794 if (dump_file && (dump_flags & TDF_DETAILS))
1795 {
1796 fprintf (dump_file, "Rejected (%d): %s: ", DECL_UID (var), msg);
1797 print_generic_expr (dump_file, var, 0);
1798 fprintf (dump_file, "\n");
1799 }
1800 }
1801
1802 /* Return true if VAR is a candidate for SRA. */
1803
1804 static bool
1805 maybe_add_sra_candidate (tree var)
1806 {
1807 tree type = TREE_TYPE (var);
1808 const char *msg;
1809 tree_node **slot;
1810
1811 if (!AGGREGATE_TYPE_P (type))
1812 {
1813 reject (var, "not aggregate");
1814 return false;
1815 }
1816 if (needs_to_live_in_memory (var))
1817 {
1818 reject (var, "needs to live in memory");
1819 return false;
1820 }
1821 if (TREE_THIS_VOLATILE (var))
1822 {
1823 reject (var, "is volatile");
1824 return false;
1825 }
1826 if (!COMPLETE_TYPE_P (type))
1827 {
1828 reject (var, "has incomplete type");
1829 return false;
1830 }
1831 if (!tree_fits_uhwi_p (TYPE_SIZE (type)))
1832 {
1833 reject (var, "type size not fixed");
1834 return false;
1835 }
1836 if (tree_to_uhwi (TYPE_SIZE (type)) == 0)
1837 {
1838 reject (var, "type size is zero");
1839 return false;
1840 }
1841 if (type_internals_preclude_sra_p (type, &msg))
1842 {
1843 reject (var, msg);
1844 return false;
1845 }
1846 if (/* Fix for PR 41089. tree-stdarg.c needs to have va_lists intact but
1847 we also want to schedule it rather late. Thus we ignore it in
1848 the early pass. */
1849 (sra_mode == SRA_MODE_EARLY_INTRA
1850 && is_va_list_type (type)))
1851 {
1852 reject (var, "is va_list");
1853 return false;
1854 }
1855
1856 bitmap_set_bit (candidate_bitmap, DECL_UID (var));
1857 slot = candidates->find_slot_with_hash (var, DECL_UID (var), INSERT);
1858 *slot = var;
1859
1860 if (dump_file && (dump_flags & TDF_DETAILS))
1861 {
1862 fprintf (dump_file, "Candidate (%d): ", DECL_UID (var));
1863 print_generic_expr (dump_file, var, 0);
1864 fprintf (dump_file, "\n");
1865 }
1866
1867 return true;
1868 }
1869
1870 /* The very first phase of intraprocedural SRA. It marks in candidate_bitmap
1871 those with type which is suitable for scalarization. */
1872
1873 static bool
1874 find_var_candidates (void)
1875 {
1876 tree var, parm;
1877 unsigned int i;
1878 bool ret = false;
1879
1880 for (parm = DECL_ARGUMENTS (current_function_decl);
1881 parm;
1882 parm = DECL_CHAIN (parm))
1883 ret |= maybe_add_sra_candidate (parm);
1884
1885 FOR_EACH_LOCAL_DECL (cfun, i, var)
1886 {
1887 if (TREE_CODE (var) != VAR_DECL)
1888 continue;
1889
1890 ret |= maybe_add_sra_candidate (var);
1891 }
1892
1893 return ret;
1894 }
1895
1896 /* Sort all accesses for the given variable, check for partial overlaps and
1897 return NULL if there are any. If there are none, pick a representative for
1898 each combination of offset and size and create a linked list out of them.
1899 Return the pointer to the first representative and make sure it is the first
1900 one in the vector of accesses. */
1901
1902 static struct access *
1903 sort_and_splice_var_accesses (tree var)
1904 {
1905 int i, j, access_count;
1906 struct access *res, **prev_acc_ptr = &res;
1907 vec<access_p> *access_vec;
1908 bool first = true;
1909 HOST_WIDE_INT low = -1, high = 0;
1910
1911 access_vec = get_base_access_vector (var);
1912 if (!access_vec)
1913 return NULL;
1914 access_count = access_vec->length ();
1915
1916 /* Sort by <OFFSET, SIZE>. */
1917 access_vec->qsort (compare_access_positions);
1918
1919 i = 0;
1920 while (i < access_count)
1921 {
1922 struct access *access = (*access_vec)[i];
1923 bool grp_write = access->write;
1924 bool grp_read = !access->write;
1925 bool grp_scalar_write = access->write
1926 && is_gimple_reg_type (access->type);
1927 bool grp_scalar_read = !access->write
1928 && is_gimple_reg_type (access->type);
1929 bool grp_assignment_read = access->grp_assignment_read;
1930 bool grp_assignment_write = access->grp_assignment_write;
1931 bool multiple_scalar_reads = false;
1932 bool total_scalarization = access->grp_total_scalarization;
1933 bool grp_partial_lhs = access->grp_partial_lhs;
1934 bool first_scalar = is_gimple_reg_type (access->type);
1935 bool unscalarizable_region = access->grp_unscalarizable_region;
1936
1937 if (first || access->offset >= high)
1938 {
1939 first = false;
1940 low = access->offset;
1941 high = access->offset + access->size;
1942 }
1943 else if (access->offset > low && access->offset + access->size > high)
1944 return NULL;
1945 else
1946 gcc_assert (access->offset >= low
1947 && access->offset + access->size <= high);
1948
1949 j = i + 1;
1950 while (j < access_count)
1951 {
1952 struct access *ac2 = (*access_vec)[j];
1953 if (ac2->offset != access->offset || ac2->size != access->size)
1954 break;
1955 if (ac2->write)
1956 {
1957 grp_write = true;
1958 grp_scalar_write = (grp_scalar_write
1959 || is_gimple_reg_type (ac2->type));
1960 }
1961 else
1962 {
1963 grp_read = true;
1964 if (is_gimple_reg_type (ac2->type))
1965 {
1966 if (grp_scalar_read)
1967 multiple_scalar_reads = true;
1968 else
1969 grp_scalar_read = true;
1970 }
1971 }
1972 grp_assignment_read |= ac2->grp_assignment_read;
1973 grp_assignment_write |= ac2->grp_assignment_write;
1974 grp_partial_lhs |= ac2->grp_partial_lhs;
1975 unscalarizable_region |= ac2->grp_unscalarizable_region;
1976 total_scalarization |= ac2->grp_total_scalarization;
1977 relink_to_new_repr (access, ac2);
1978
1979 /* If there are both aggregate-type and scalar-type accesses with
1980 this combination of size and offset, the comparison function
1981 should have put the scalars first. */
1982 gcc_assert (first_scalar || !is_gimple_reg_type (ac2->type));
1983 ac2->group_representative = access;
1984 j++;
1985 }
1986
1987 i = j;
1988
1989 access->group_representative = access;
1990 access->grp_write = grp_write;
1991 access->grp_read = grp_read;
1992 access->grp_scalar_read = grp_scalar_read;
1993 access->grp_scalar_write = grp_scalar_write;
1994 access->grp_assignment_read = grp_assignment_read;
1995 access->grp_assignment_write = grp_assignment_write;
1996 access->grp_hint = multiple_scalar_reads || total_scalarization;
1997 access->grp_total_scalarization = total_scalarization;
1998 access->grp_partial_lhs = grp_partial_lhs;
1999 access->grp_unscalarizable_region = unscalarizable_region;
2000 if (access->first_link)
2001 add_access_to_work_queue (access);
2002
2003 *prev_acc_ptr = access;
2004 prev_acc_ptr = &access->next_grp;
2005 }
2006
2007 gcc_assert (res == (*access_vec)[0]);
2008 return res;
2009 }
2010
2011 /* Create a variable for the given ACCESS which determines the type, name and a
2012 few other properties. Return the variable declaration and store it also to
2013 ACCESS->replacement. */
2014
2015 static tree
2016 create_access_replacement (struct access *access)
2017 {
2018 tree repl;
2019
2020 if (access->grp_to_be_debug_replaced)
2021 {
2022 repl = create_tmp_var_raw (access->type);
2023 DECL_CONTEXT (repl) = current_function_decl;
2024 }
2025 else
2026 /* Drop any special alignment on the type if it's not on the main
2027 variant. This avoids issues with weirdo ABIs like AAPCS. */
2028 repl = create_tmp_var (build_qualified_type
2029 (TYPE_MAIN_VARIANT (access->type),
2030 TYPE_QUALS (access->type)), "SR");
2031 if (TREE_CODE (access->type) == COMPLEX_TYPE
2032 || TREE_CODE (access->type) == VECTOR_TYPE)
2033 {
2034 if (!access->grp_partial_lhs)
2035 DECL_GIMPLE_REG_P (repl) = 1;
2036 }
2037 else if (access->grp_partial_lhs
2038 && is_gimple_reg_type (access->type))
2039 TREE_ADDRESSABLE (repl) = 1;
2040
2041 DECL_SOURCE_LOCATION (repl) = DECL_SOURCE_LOCATION (access->base);
2042 DECL_ARTIFICIAL (repl) = 1;
2043 DECL_IGNORED_P (repl) = DECL_IGNORED_P (access->base);
2044
2045 if (DECL_NAME (access->base)
2046 && !DECL_IGNORED_P (access->base)
2047 && !DECL_ARTIFICIAL (access->base))
2048 {
2049 char *pretty_name = make_fancy_name (access->expr);
2050 tree debug_expr = unshare_expr_without_location (access->expr), d;
2051 bool fail = false;
2052
2053 DECL_NAME (repl) = get_identifier (pretty_name);
2054 obstack_free (&name_obstack, pretty_name);
2055
2056 /* Get rid of any SSA_NAMEs embedded in debug_expr,
2057 as DECL_DEBUG_EXPR isn't considered when looking for still
2058 used SSA_NAMEs and thus they could be freed. All debug info
2059 generation cares is whether something is constant or variable
2060 and that get_ref_base_and_extent works properly on the
2061 expression. It cannot handle accesses at a non-constant offset
2062 though, so just give up in those cases. */
2063 for (d = debug_expr;
2064 !fail && (handled_component_p (d) || TREE_CODE (d) == MEM_REF);
2065 d = TREE_OPERAND (d, 0))
2066 switch (TREE_CODE (d))
2067 {
2068 case ARRAY_REF:
2069 case ARRAY_RANGE_REF:
2070 if (TREE_OPERAND (d, 1)
2071 && TREE_CODE (TREE_OPERAND (d, 1)) != INTEGER_CST)
2072 fail = true;
2073 if (TREE_OPERAND (d, 3)
2074 && TREE_CODE (TREE_OPERAND (d, 3)) != INTEGER_CST)
2075 fail = true;
2076 /* FALLTHRU */
2077 case COMPONENT_REF:
2078 if (TREE_OPERAND (d, 2)
2079 && TREE_CODE (TREE_OPERAND (d, 2)) != INTEGER_CST)
2080 fail = true;
2081 break;
2082 case MEM_REF:
2083 if (TREE_CODE (TREE_OPERAND (d, 0)) != ADDR_EXPR)
2084 fail = true;
2085 else
2086 d = TREE_OPERAND (d, 0);
2087 break;
2088 default:
2089 break;
2090 }
2091 if (!fail)
2092 {
2093 SET_DECL_DEBUG_EXPR (repl, debug_expr);
2094 DECL_HAS_DEBUG_EXPR_P (repl) = 1;
2095 }
2096 if (access->grp_no_warning)
2097 TREE_NO_WARNING (repl) = 1;
2098 else
2099 TREE_NO_WARNING (repl) = TREE_NO_WARNING (access->base);
2100 }
2101 else
2102 TREE_NO_WARNING (repl) = 1;
2103
2104 if (dump_file)
2105 {
2106 if (access->grp_to_be_debug_replaced)
2107 {
2108 fprintf (dump_file, "Created a debug-only replacement for ");
2109 print_generic_expr (dump_file, access->base, 0);
2110 fprintf (dump_file, " offset: %u, size: %u\n",
2111 (unsigned) access->offset, (unsigned) access->size);
2112 }
2113 else
2114 {
2115 fprintf (dump_file, "Created a replacement for ");
2116 print_generic_expr (dump_file, access->base, 0);
2117 fprintf (dump_file, " offset: %u, size: %u: ",
2118 (unsigned) access->offset, (unsigned) access->size);
2119 print_generic_expr (dump_file, repl, 0);
2120 fprintf (dump_file, "\n");
2121 }
2122 }
2123 sra_stats.replacements++;
2124
2125 return repl;
2126 }
2127
2128 /* Return ACCESS scalar replacement, create it if it does not exist yet. */
2129
2130 static inline tree
2131 get_access_replacement (struct access *access)
2132 {
2133 gcc_checking_assert (access->replacement_decl);
2134 return access->replacement_decl;
2135 }
2136
2137
2138 /* Build a subtree of accesses rooted in *ACCESS, and move the pointer in the
2139 linked list along the way. Stop when *ACCESS is NULL or the access pointed
2140 to it is not "within" the root. Return false iff some accesses partially
2141 overlap. */
2142
2143 static bool
2144 build_access_subtree (struct access **access)
2145 {
2146 struct access *root = *access, *last_child = NULL;
2147 HOST_WIDE_INT limit = root->offset + root->size;
2148
2149 *access = (*access)->next_grp;
2150 while (*access && (*access)->offset + (*access)->size <= limit)
2151 {
2152 if (!last_child)
2153 root->first_child = *access;
2154 else
2155 last_child->next_sibling = *access;
2156 last_child = *access;
2157
2158 if (!build_access_subtree (access))
2159 return false;
2160 }
2161
2162 if (*access && (*access)->offset < limit)
2163 return false;
2164
2165 return true;
2166 }
2167
2168 /* Build a tree of access representatives, ACCESS is the pointer to the first
2169 one, others are linked in a list by the next_grp field. Return false iff
2170 some accesses partially overlap. */
2171
2172 static bool
2173 build_access_trees (struct access *access)
2174 {
2175 while (access)
2176 {
2177 struct access *root = access;
2178
2179 if (!build_access_subtree (&access))
2180 return false;
2181 root->next_grp = access;
2182 }
2183 return true;
2184 }
2185
2186 /* Return true if expr contains some ARRAY_REFs into a variable bounded
2187 array. */
2188
2189 static bool
2190 expr_with_var_bounded_array_refs_p (tree expr)
2191 {
2192 while (handled_component_p (expr))
2193 {
2194 if (TREE_CODE (expr) == ARRAY_REF
2195 && !tree_fits_shwi_p (array_ref_low_bound (expr)))
2196 return true;
2197 expr = TREE_OPERAND (expr, 0);
2198 }
2199 return false;
2200 }
2201
2202 /* Analyze the subtree of accesses rooted in ROOT, scheduling replacements when
2203 both seeming beneficial and when ALLOW_REPLACEMENTS allows it. Also set all
2204 sorts of access flags appropriately along the way, notably always set
2205 grp_read and grp_assign_read according to MARK_READ and grp_write when
2206 MARK_WRITE is true.
2207
2208 Creating a replacement for a scalar access is considered beneficial if its
2209 grp_hint is set (this means we are either attempting total scalarization or
2210 there is more than one direct read access) or according to the following
2211 table:
2212
2213 Access written to through a scalar type (once or more times)
2214 |
2215 | Written to in an assignment statement
2216 | |
2217 | | Access read as scalar _once_
2218 | | |
2219 | | | Read in an assignment statement
2220 | | | |
2221 | | | | Scalarize Comment
2222 -----------------------------------------------------------------------------
2223 0 0 0 0 No access for the scalar
2224 0 0 0 1 No access for the scalar
2225 0 0 1 0 No Single read - won't help
2226 0 0 1 1 No The same case
2227 0 1 0 0 No access for the scalar
2228 0 1 0 1 No access for the scalar
2229 0 1 1 0 Yes s = *g; return s.i;
2230 0 1 1 1 Yes The same case as above
2231 1 0 0 0 No Won't help
2232 1 0 0 1 Yes s.i = 1; *g = s;
2233 1 0 1 0 Yes s.i = 5; g = s.i;
2234 1 0 1 1 Yes The same case as above
2235 1 1 0 0 No Won't help.
2236 1 1 0 1 Yes s.i = 1; *g = s;
2237 1 1 1 0 Yes s = *g; return s.i;
2238 1 1 1 1 Yes Any of the above yeses */
2239
2240 static bool
2241 analyze_access_subtree (struct access *root, struct access *parent,
2242 bool allow_replacements)
2243 {
2244 struct access *child;
2245 HOST_WIDE_INT limit = root->offset + root->size;
2246 HOST_WIDE_INT covered_to = root->offset;
2247 bool scalar = is_gimple_reg_type (root->type);
2248 bool hole = false, sth_created = false;
2249
2250 if (parent)
2251 {
2252 if (parent->grp_read)
2253 root->grp_read = 1;
2254 if (parent->grp_assignment_read)
2255 root->grp_assignment_read = 1;
2256 if (parent->grp_write)
2257 root->grp_write = 1;
2258 if (parent->grp_assignment_write)
2259 root->grp_assignment_write = 1;
2260 if (parent->grp_total_scalarization)
2261 root->grp_total_scalarization = 1;
2262 }
2263
2264 if (root->grp_unscalarizable_region)
2265 allow_replacements = false;
2266
2267 if (allow_replacements && expr_with_var_bounded_array_refs_p (root->expr))
2268 allow_replacements = false;
2269
2270 for (child = root->first_child; child; child = child->next_sibling)
2271 {
2272 hole |= covered_to < child->offset;
2273 sth_created |= analyze_access_subtree (child, root,
2274 allow_replacements && !scalar);
2275
2276 root->grp_unscalarized_data |= child->grp_unscalarized_data;
2277 root->grp_total_scalarization &= child->grp_total_scalarization;
2278 if (child->grp_covered)
2279 covered_to += child->size;
2280 else
2281 hole = true;
2282 }
2283
2284 if (allow_replacements && scalar && !root->first_child
2285 && (root->grp_hint
2286 || ((root->grp_scalar_read || root->grp_assignment_read)
2287 && (root->grp_scalar_write || root->grp_assignment_write))))
2288 {
2289 /* Always create access replacements that cover the whole access.
2290 For integral types this means the precision has to match.
2291 Avoid assumptions based on the integral type kind, too. */
2292 if (INTEGRAL_TYPE_P (root->type)
2293 && (TREE_CODE (root->type) != INTEGER_TYPE
2294 || TYPE_PRECISION (root->type) != root->size)
2295 /* But leave bitfield accesses alone. */
2296 && (TREE_CODE (root->expr) != COMPONENT_REF
2297 || !DECL_BIT_FIELD (TREE_OPERAND (root->expr, 1))))
2298 {
2299 tree rt = root->type;
2300 gcc_assert ((root->offset % BITS_PER_UNIT) == 0
2301 && (root->size % BITS_PER_UNIT) == 0);
2302 root->type = build_nonstandard_integer_type (root->size,
2303 TYPE_UNSIGNED (rt));
2304 root->expr = build_ref_for_offset (UNKNOWN_LOCATION,
2305 root->base, root->offset,
2306 root->type, NULL, false);
2307
2308 if (dump_file && (dump_flags & TDF_DETAILS))
2309 {
2310 fprintf (dump_file, "Changing the type of a replacement for ");
2311 print_generic_expr (dump_file, root->base, 0);
2312 fprintf (dump_file, " offset: %u, size: %u ",
2313 (unsigned) root->offset, (unsigned) root->size);
2314 fprintf (dump_file, " to an integer.\n");
2315 }
2316 }
2317
2318 root->grp_to_be_replaced = 1;
2319 root->replacement_decl = create_access_replacement (root);
2320 sth_created = true;
2321 hole = false;
2322 }
2323 else
2324 {
2325 if (allow_replacements
2326 && scalar && !root->first_child
2327 && (root->grp_scalar_write || root->grp_assignment_write)
2328 && !bitmap_bit_p (cannot_scalarize_away_bitmap,
2329 DECL_UID (root->base)))
2330 {
2331 gcc_checking_assert (!root->grp_scalar_read
2332 && !root->grp_assignment_read);
2333 sth_created = true;
2334 if (MAY_HAVE_DEBUG_STMTS)
2335 {
2336 root->grp_to_be_debug_replaced = 1;
2337 root->replacement_decl = create_access_replacement (root);
2338 }
2339 }
2340
2341 if (covered_to < limit)
2342 hole = true;
2343 if (scalar)
2344 root->grp_total_scalarization = 0;
2345 }
2346
2347 if (!hole || root->grp_total_scalarization)
2348 root->grp_covered = 1;
2349 else if (root->grp_write || TREE_CODE (root->base) == PARM_DECL)
2350 root->grp_unscalarized_data = 1; /* not covered and written to */
2351 return sth_created;
2352 }
2353
2354 /* Analyze all access trees linked by next_grp by the means of
2355 analyze_access_subtree. */
2356 static bool
2357 analyze_access_trees (struct access *access)
2358 {
2359 bool ret = false;
2360
2361 while (access)
2362 {
2363 if (analyze_access_subtree (access, NULL, true))
2364 ret = true;
2365 access = access->next_grp;
2366 }
2367
2368 return ret;
2369 }
2370
2371 /* Return true iff a potential new child of LACC at offset OFFSET and with size
2372 SIZE would conflict with an already existing one. If exactly such a child
2373 already exists in LACC, store a pointer to it in EXACT_MATCH. */
2374
2375 static bool
2376 child_would_conflict_in_lacc (struct access *lacc, HOST_WIDE_INT norm_offset,
2377 HOST_WIDE_INT size, struct access **exact_match)
2378 {
2379 struct access *child;
2380
2381 for (child = lacc->first_child; child; child = child->next_sibling)
2382 {
2383 if (child->offset == norm_offset && child->size == size)
2384 {
2385 *exact_match = child;
2386 return true;
2387 }
2388
2389 if (child->offset < norm_offset + size
2390 && child->offset + child->size > norm_offset)
2391 return true;
2392 }
2393
2394 return false;
2395 }
2396
2397 /* Create a new child access of PARENT, with all properties just like MODEL
2398 except for its offset and with its grp_write false and grp_read true.
2399 Return the new access or NULL if it cannot be created. Note that this access
2400 is created long after all splicing and sorting, it's not located in any
2401 access vector and is automatically a representative of its group. */
2402
2403 static struct access *
2404 create_artificial_child_access (struct access *parent, struct access *model,
2405 HOST_WIDE_INT new_offset)
2406 {
2407 struct access **child;
2408 tree expr = parent->base;
2409
2410 gcc_assert (!model->grp_unscalarizable_region);
2411
2412 struct access *access = new struct access ();
2413 memset (access, 0, sizeof (struct access));
2414 if (!build_user_friendly_ref_for_offset (&expr, TREE_TYPE (expr), new_offset,
2415 model->type))
2416 {
2417 access->grp_no_warning = true;
2418 expr = build_ref_for_model (EXPR_LOCATION (parent->base), parent->base,
2419 new_offset, model, NULL, false);
2420 }
2421
2422 access->base = parent->base;
2423 access->expr = expr;
2424 access->offset = new_offset;
2425 access->size = model->size;
2426 access->type = model->type;
2427 access->grp_write = true;
2428 access->grp_read = false;
2429
2430 child = &parent->first_child;
2431 while (*child && (*child)->offset < new_offset)
2432 child = &(*child)->next_sibling;
2433
2434 access->next_sibling = *child;
2435 *child = access;
2436
2437 return access;
2438 }
2439
2440
2441 /* Propagate all subaccesses of RACC across an assignment link to LACC. Return
2442 true if any new subaccess was created. Additionally, if RACC is a scalar
2443 access but LACC is not, change the type of the latter, if possible. */
2444
2445 static bool
2446 propagate_subaccesses_across_link (struct access *lacc, struct access *racc)
2447 {
2448 struct access *rchild;
2449 HOST_WIDE_INT norm_delta = lacc->offset - racc->offset;
2450 bool ret = false;
2451
2452 if (is_gimple_reg_type (lacc->type)
2453 || lacc->grp_unscalarizable_region
2454 || racc->grp_unscalarizable_region)
2455 return false;
2456
2457 if (is_gimple_reg_type (racc->type))
2458 {
2459 if (!lacc->first_child && !racc->first_child)
2460 {
2461 tree t = lacc->base;
2462
2463 lacc->type = racc->type;
2464 if (build_user_friendly_ref_for_offset (&t, TREE_TYPE (t),
2465 lacc->offset, racc->type))
2466 lacc->expr = t;
2467 else
2468 {
2469 lacc->expr = build_ref_for_model (EXPR_LOCATION (lacc->base),
2470 lacc->base, lacc->offset,
2471 racc, NULL, false);
2472 lacc->grp_no_warning = true;
2473 }
2474 }
2475 return false;
2476 }
2477
2478 for (rchild = racc->first_child; rchild; rchild = rchild->next_sibling)
2479 {
2480 struct access *new_acc = NULL;
2481 HOST_WIDE_INT norm_offset = rchild->offset + norm_delta;
2482
2483 if (rchild->grp_unscalarizable_region)
2484 continue;
2485
2486 if (child_would_conflict_in_lacc (lacc, norm_offset, rchild->size,
2487 &new_acc))
2488 {
2489 if (new_acc)
2490 {
2491 rchild->grp_hint = 1;
2492 new_acc->grp_hint |= new_acc->grp_read;
2493 if (rchild->first_child)
2494 ret |= propagate_subaccesses_across_link (new_acc, rchild);
2495 }
2496 continue;
2497 }
2498
2499 rchild->grp_hint = 1;
2500 new_acc = create_artificial_child_access (lacc, rchild, norm_offset);
2501 if (new_acc)
2502 {
2503 ret = true;
2504 if (racc->first_child)
2505 propagate_subaccesses_across_link (new_acc, rchild);
2506 }
2507 }
2508
2509 return ret;
2510 }
2511
2512 /* Propagate all subaccesses across assignment links. */
2513
2514 static void
2515 propagate_all_subaccesses (void)
2516 {
2517 while (work_queue_head)
2518 {
2519 struct access *racc = pop_access_from_work_queue ();
2520 struct assign_link *link;
2521
2522 gcc_assert (racc->first_link);
2523
2524 for (link = racc->first_link; link; link = link->next)
2525 {
2526 struct access *lacc = link->lacc;
2527
2528 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (lacc->base)))
2529 continue;
2530 lacc = lacc->group_representative;
2531 if (propagate_subaccesses_across_link (lacc, racc)
2532 && lacc->first_link)
2533 add_access_to_work_queue (lacc);
2534 }
2535 }
2536 }
2537
2538 /* Go through all accesses collected throughout the (intraprocedural) analysis
2539 stage, exclude overlapping ones, identify representatives and build trees
2540 out of them, making decisions about scalarization on the way. Return true
2541 iff there are any to-be-scalarized variables after this stage. */
2542
2543 static bool
2544 analyze_all_variable_accesses (void)
2545 {
2546 int res = 0;
2547 bitmap tmp = BITMAP_ALLOC (NULL);
2548 bitmap_iterator bi;
2549 unsigned i;
2550 unsigned max_scalarization_size
2551 = (optimize_function_for_size_p (cfun)
2552 ? PARAM_VALUE (PARAM_SRA_MAX_SCALARIZATION_SIZE_SIZE)
2553 : PARAM_VALUE (PARAM_SRA_MAX_SCALARIZATION_SIZE_SPEED))
2554 * BITS_PER_UNIT;
2555
2556 EXECUTE_IF_SET_IN_BITMAP (candidate_bitmap, 0, i, bi)
2557 if (bitmap_bit_p (should_scalarize_away_bitmap, i)
2558 && !bitmap_bit_p (cannot_scalarize_away_bitmap, i))
2559 {
2560 tree var = candidate (i);
2561
2562 if (TREE_CODE (var) == VAR_DECL
2563 && type_consists_of_records_p (TREE_TYPE (var)))
2564 {
2565 if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
2566 <= max_scalarization_size)
2567 {
2568 completely_scalarize_var (var);
2569 if (dump_file && (dump_flags & TDF_DETAILS))
2570 {
2571 fprintf (dump_file, "Will attempt to totally scalarize ");
2572 print_generic_expr (dump_file, var, 0);
2573 fprintf (dump_file, " (UID: %u): \n", DECL_UID (var));
2574 }
2575 }
2576 else if (dump_file && (dump_flags & TDF_DETAILS))
2577 {
2578 fprintf (dump_file, "Too big to totally scalarize: ");
2579 print_generic_expr (dump_file, var, 0);
2580 fprintf (dump_file, " (UID: %u)\n", DECL_UID (var));
2581 }
2582 }
2583 }
2584
2585 bitmap_copy (tmp, candidate_bitmap);
2586 EXECUTE_IF_SET_IN_BITMAP (tmp, 0, i, bi)
2587 {
2588 tree var = candidate (i);
2589 struct access *access;
2590
2591 access = sort_and_splice_var_accesses (var);
2592 if (!access || !build_access_trees (access))
2593 disqualify_candidate (var,
2594 "No or inhibitingly overlapping accesses.");
2595 }
2596
2597 propagate_all_subaccesses ();
2598
2599 bitmap_copy (tmp, candidate_bitmap);
2600 EXECUTE_IF_SET_IN_BITMAP (tmp, 0, i, bi)
2601 {
2602 tree var = candidate (i);
2603 struct access *access = get_first_repr_for_decl (var);
2604
2605 if (analyze_access_trees (access))
2606 {
2607 res++;
2608 if (dump_file && (dump_flags & TDF_DETAILS))
2609 {
2610 fprintf (dump_file, "\nAccess trees for ");
2611 print_generic_expr (dump_file, var, 0);
2612 fprintf (dump_file, " (UID: %u): \n", DECL_UID (var));
2613 dump_access_tree (dump_file, access);
2614 fprintf (dump_file, "\n");
2615 }
2616 }
2617 else
2618 disqualify_candidate (var, "No scalar replacements to be created.");
2619 }
2620
2621 BITMAP_FREE (tmp);
2622
2623 if (res)
2624 {
2625 statistics_counter_event (cfun, "Scalarized aggregates", res);
2626 return true;
2627 }
2628 else
2629 return false;
2630 }
2631
2632 /* Generate statements copying scalar replacements of accesses within a subtree
2633 into or out of AGG. ACCESS, all its children, siblings and their children
2634 are to be processed. AGG is an aggregate type expression (can be a
2635 declaration but does not have to be, it can for example also be a mem_ref or
2636 a series of handled components). TOP_OFFSET is the offset of the processed
2637 subtree which has to be subtracted from offsets of individual accesses to
2638 get corresponding offsets for AGG. If CHUNK_SIZE is non-null, copy only
2639 replacements in the interval <start_offset, start_offset + chunk_size>,
2640 otherwise copy all. GSI is a statement iterator used to place the new
2641 statements. WRITE should be true when the statements should write from AGG
2642 to the replacement and false if vice versa. if INSERT_AFTER is true, new
2643 statements will be added after the current statement in GSI, they will be
2644 added before the statement otherwise. */
2645
2646 static void
2647 generate_subtree_copies (struct access *access, tree agg,
2648 HOST_WIDE_INT top_offset,
2649 HOST_WIDE_INT start_offset, HOST_WIDE_INT chunk_size,
2650 gimple_stmt_iterator *gsi, bool write,
2651 bool insert_after, location_t loc)
2652 {
2653 do
2654 {
2655 if (chunk_size && access->offset >= start_offset + chunk_size)
2656 return;
2657
2658 if (access->grp_to_be_replaced
2659 && (chunk_size == 0
2660 || access->offset + access->size > start_offset))
2661 {
2662 tree expr, repl = get_access_replacement (access);
2663 gassign *stmt;
2664
2665 expr = build_ref_for_model (loc, agg, access->offset - top_offset,
2666 access, gsi, insert_after);
2667
2668 if (write)
2669 {
2670 if (access->grp_partial_lhs)
2671 expr = force_gimple_operand_gsi (gsi, expr, true, NULL_TREE,
2672 !insert_after,
2673 insert_after ? GSI_NEW_STMT
2674 : GSI_SAME_STMT);
2675 stmt = gimple_build_assign (repl, expr);
2676 }
2677 else
2678 {
2679 TREE_NO_WARNING (repl) = 1;
2680 if (access->grp_partial_lhs)
2681 repl = force_gimple_operand_gsi (gsi, repl, true, NULL_TREE,
2682 !insert_after,
2683 insert_after ? GSI_NEW_STMT
2684 : GSI_SAME_STMT);
2685 stmt = gimple_build_assign (expr, repl);
2686 }
2687 gimple_set_location (stmt, loc);
2688
2689 if (insert_after)
2690 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2691 else
2692 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2693 update_stmt (stmt);
2694 sra_stats.subtree_copies++;
2695 }
2696 else if (write
2697 && access->grp_to_be_debug_replaced
2698 && (chunk_size == 0
2699 || access->offset + access->size > start_offset))
2700 {
2701 gdebug *ds;
2702 tree drhs = build_debug_ref_for_model (loc, agg,
2703 access->offset - top_offset,
2704 access);
2705 ds = gimple_build_debug_bind (get_access_replacement (access),
2706 drhs, gsi_stmt (*gsi));
2707 if (insert_after)
2708 gsi_insert_after (gsi, ds, GSI_NEW_STMT);
2709 else
2710 gsi_insert_before (gsi, ds, GSI_SAME_STMT);
2711 }
2712
2713 if (access->first_child)
2714 generate_subtree_copies (access->first_child, agg, top_offset,
2715 start_offset, chunk_size, gsi,
2716 write, insert_after, loc);
2717
2718 access = access->next_sibling;
2719 }
2720 while (access);
2721 }
2722
2723 /* Assign zero to all scalar replacements in an access subtree. ACCESS is the
2724 the root of the subtree to be processed. GSI is the statement iterator used
2725 for inserting statements which are added after the current statement if
2726 INSERT_AFTER is true or before it otherwise. */
2727
2728 static void
2729 init_subtree_with_zero (struct access *access, gimple_stmt_iterator *gsi,
2730 bool insert_after, location_t loc)
2731
2732 {
2733 struct access *child;
2734
2735 if (access->grp_to_be_replaced)
2736 {
2737 gassign *stmt;
2738
2739 stmt = gimple_build_assign (get_access_replacement (access),
2740 build_zero_cst (access->type));
2741 if (insert_after)
2742 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2743 else
2744 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2745 update_stmt (stmt);
2746 gimple_set_location (stmt, loc);
2747 }
2748 else if (access->grp_to_be_debug_replaced)
2749 {
2750 gdebug *ds
2751 = gimple_build_debug_bind (get_access_replacement (access),
2752 build_zero_cst (access->type),
2753 gsi_stmt (*gsi));
2754 if (insert_after)
2755 gsi_insert_after (gsi, ds, GSI_NEW_STMT);
2756 else
2757 gsi_insert_before (gsi, ds, GSI_SAME_STMT);
2758 }
2759
2760 for (child = access->first_child; child; child = child->next_sibling)
2761 init_subtree_with_zero (child, gsi, insert_after, loc);
2762 }
2763
2764 /* Clobber all scalar replacements in an access subtree. ACCESS is the the
2765 root of the subtree to be processed. GSI is the statement iterator used
2766 for inserting statements which are added after the current statement if
2767 INSERT_AFTER is true or before it otherwise. */
2768
2769 static void
2770 clobber_subtree (struct access *access, gimple_stmt_iterator *gsi,
2771 bool insert_after, location_t loc)
2772
2773 {
2774 struct access *child;
2775
2776 if (access->grp_to_be_replaced)
2777 {
2778 tree rep = get_access_replacement (access);
2779 tree clobber = build_constructor (access->type, NULL);
2780 TREE_THIS_VOLATILE (clobber) = 1;
2781 gimple stmt = gimple_build_assign (rep, clobber);
2782
2783 if (insert_after)
2784 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2785 else
2786 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2787 update_stmt (stmt);
2788 gimple_set_location (stmt, loc);
2789 }
2790
2791 for (child = access->first_child; child; child = child->next_sibling)
2792 clobber_subtree (child, gsi, insert_after, loc);
2793 }
2794
2795 /* Search for an access representative for the given expression EXPR and
2796 return it or NULL if it cannot be found. */
2797
2798 static struct access *
2799 get_access_for_expr (tree expr)
2800 {
2801 HOST_WIDE_INT offset, size, max_size;
2802 tree base;
2803
2804 /* FIXME: This should not be necessary but Ada produces V_C_Es with a type of
2805 a different size than the size of its argument and we need the latter
2806 one. */
2807 if (TREE_CODE (expr) == VIEW_CONVERT_EXPR)
2808 expr = TREE_OPERAND (expr, 0);
2809
2810 base = get_ref_base_and_extent (expr, &offset, &size, &max_size);
2811 if (max_size == -1 || !DECL_P (base))
2812 return NULL;
2813
2814 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (base)))
2815 return NULL;
2816
2817 return get_var_base_offset_size_access (base, offset, max_size);
2818 }
2819
2820 /* Replace the expression EXPR with a scalar replacement if there is one and
2821 generate other statements to do type conversion or subtree copying if
2822 necessary. GSI is used to place newly created statements, WRITE is true if
2823 the expression is being written to (it is on a LHS of a statement or output
2824 in an assembly statement). */
2825
2826 static bool
2827 sra_modify_expr (tree *expr, gimple_stmt_iterator *gsi, bool write)
2828 {
2829 location_t loc;
2830 struct access *access;
2831 tree type, bfr, orig_expr;
2832
2833 if (TREE_CODE (*expr) == BIT_FIELD_REF)
2834 {
2835 bfr = *expr;
2836 expr = &TREE_OPERAND (*expr, 0);
2837 }
2838 else
2839 bfr = NULL_TREE;
2840
2841 if (TREE_CODE (*expr) == REALPART_EXPR || TREE_CODE (*expr) == IMAGPART_EXPR)
2842 expr = &TREE_OPERAND (*expr, 0);
2843 access = get_access_for_expr (*expr);
2844 if (!access)
2845 return false;
2846 type = TREE_TYPE (*expr);
2847 orig_expr = *expr;
2848
2849 loc = gimple_location (gsi_stmt (*gsi));
2850 gimple_stmt_iterator alt_gsi = gsi_none ();
2851 if (write && stmt_ends_bb_p (gsi_stmt (*gsi)))
2852 {
2853 alt_gsi = gsi_start_edge (single_non_eh_succ (gsi_bb (*gsi)));
2854 gsi = &alt_gsi;
2855 }
2856
2857 if (access->grp_to_be_replaced)
2858 {
2859 tree repl = get_access_replacement (access);
2860 /* If we replace a non-register typed access simply use the original
2861 access expression to extract the scalar component afterwards.
2862 This happens if scalarizing a function return value or parameter
2863 like in gcc.c-torture/execute/20041124-1.c, 20050316-1.c and
2864 gcc.c-torture/compile/20011217-1.c.
2865
2866 We also want to use this when accessing a complex or vector which can
2867 be accessed as a different type too, potentially creating a need for
2868 type conversion (see PR42196) and when scalarized unions are involved
2869 in assembler statements (see PR42398). */
2870 if (!useless_type_conversion_p (type, access->type))
2871 {
2872 tree ref;
2873
2874 ref = build_ref_for_model (loc, orig_expr, 0, access, gsi, false);
2875
2876 if (write)
2877 {
2878 gassign *stmt;
2879
2880 if (access->grp_partial_lhs)
2881 ref = force_gimple_operand_gsi (gsi, ref, true, NULL_TREE,
2882 false, GSI_NEW_STMT);
2883 stmt = gimple_build_assign (repl, ref);
2884 gimple_set_location (stmt, loc);
2885 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2886 }
2887 else
2888 {
2889 gassign *stmt;
2890
2891 if (access->grp_partial_lhs)
2892 repl = force_gimple_operand_gsi (gsi, repl, true, NULL_TREE,
2893 true, GSI_SAME_STMT);
2894 stmt = gimple_build_assign (ref, repl);
2895 gimple_set_location (stmt, loc);
2896 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2897 }
2898 }
2899 else
2900 *expr = repl;
2901 sra_stats.exprs++;
2902 }
2903 else if (write && access->grp_to_be_debug_replaced)
2904 {
2905 gdebug *ds = gimple_build_debug_bind (get_access_replacement (access),
2906 NULL_TREE,
2907 gsi_stmt (*gsi));
2908 gsi_insert_after (gsi, ds, GSI_NEW_STMT);
2909 }
2910
2911 if (access->first_child)
2912 {
2913 HOST_WIDE_INT start_offset, chunk_size;
2914 if (bfr
2915 && tree_fits_uhwi_p (TREE_OPERAND (bfr, 1))
2916 && tree_fits_uhwi_p (TREE_OPERAND (bfr, 2)))
2917 {
2918 chunk_size = tree_to_uhwi (TREE_OPERAND (bfr, 1));
2919 start_offset = access->offset
2920 + tree_to_uhwi (TREE_OPERAND (bfr, 2));
2921 }
2922 else
2923 start_offset = chunk_size = 0;
2924
2925 generate_subtree_copies (access->first_child, orig_expr, access->offset,
2926 start_offset, chunk_size, gsi, write, write,
2927 loc);
2928 }
2929 return true;
2930 }
2931
2932 /* Where scalar replacements of the RHS have been written to when a replacement
2933 of a LHS of an assigments cannot be direclty loaded from a replacement of
2934 the RHS. */
2935 enum unscalarized_data_handling { SRA_UDH_NONE, /* Nothing done so far. */
2936 SRA_UDH_RIGHT, /* Data flushed to the RHS. */
2937 SRA_UDH_LEFT }; /* Data flushed to the LHS. */
2938
2939 struct subreplacement_assignment_data
2940 {
2941 /* Offset of the access representing the lhs of the assignment. */
2942 HOST_WIDE_INT left_offset;
2943
2944 /* LHS and RHS of the original assignment. */
2945 tree assignment_lhs, assignment_rhs;
2946
2947 /* Access representing the rhs of the whole assignment. */
2948 struct access *top_racc;
2949
2950 /* Stmt iterator used for statement insertions after the original assignment.
2951 It points to the main GSI used to traverse a BB during function body
2952 modification. */
2953 gimple_stmt_iterator *new_gsi;
2954
2955 /* Stmt iterator used for statement insertions before the original
2956 assignment. Keeps on pointing to the original statement. */
2957 gimple_stmt_iterator old_gsi;
2958
2959 /* Location of the assignment. */
2960 location_t loc;
2961
2962 /* Keeps the information whether we have needed to refresh replacements of
2963 the LHS and from which side of the assignments this takes place. */
2964 enum unscalarized_data_handling refreshed;
2965 };
2966
2967 /* Store all replacements in the access tree rooted in TOP_RACC either to their
2968 base aggregate if there are unscalarized data or directly to LHS of the
2969 statement that is pointed to by GSI otherwise. */
2970
2971 static void
2972 handle_unscalarized_data_in_subtree (struct subreplacement_assignment_data *sad)
2973 {
2974 tree src;
2975 if (sad->top_racc->grp_unscalarized_data)
2976 {
2977 src = sad->assignment_rhs;
2978 sad->refreshed = SRA_UDH_RIGHT;
2979 }
2980 else
2981 {
2982 src = sad->assignment_lhs;
2983 sad->refreshed = SRA_UDH_LEFT;
2984 }
2985 generate_subtree_copies (sad->top_racc->first_child, src,
2986 sad->top_racc->offset, 0, 0,
2987 &sad->old_gsi, false, false, sad->loc);
2988 }
2989
2990 /* Try to generate statements to load all sub-replacements in an access subtree
2991 formed by children of LACC from scalar replacements in the SAD->top_racc
2992 subtree. If that is not possible, refresh the SAD->top_racc base aggregate
2993 and load the accesses from it. */
2994
2995 static void
2996 load_assign_lhs_subreplacements (struct access *lacc,
2997 struct subreplacement_assignment_data *sad)
2998 {
2999 for (lacc = lacc->first_child; lacc; lacc = lacc->next_sibling)
3000 {
3001 HOST_WIDE_INT offset;
3002 offset = lacc->offset - sad->left_offset + sad->top_racc->offset;
3003
3004 if (lacc->grp_to_be_replaced)
3005 {
3006 struct access *racc;
3007 gassign *stmt;
3008 tree rhs;
3009
3010 racc = find_access_in_subtree (sad->top_racc, offset, lacc->size);
3011 if (racc && racc->grp_to_be_replaced)
3012 {
3013 rhs = get_access_replacement (racc);
3014 if (!useless_type_conversion_p (lacc->type, racc->type))
3015 rhs = fold_build1_loc (sad->loc, VIEW_CONVERT_EXPR,
3016 lacc->type, rhs);
3017
3018 if (racc->grp_partial_lhs && lacc->grp_partial_lhs)
3019 rhs = force_gimple_operand_gsi (&sad->old_gsi, rhs, true,
3020 NULL_TREE, true, GSI_SAME_STMT);
3021 }
3022 else
3023 {
3024 /* No suitable access on the right hand side, need to load from
3025 the aggregate. See if we have to update it first... */
3026 if (sad->refreshed == SRA_UDH_NONE)
3027 handle_unscalarized_data_in_subtree (sad);
3028
3029 if (sad->refreshed == SRA_UDH_LEFT)
3030 rhs = build_ref_for_model (sad->loc, sad->assignment_lhs,
3031 lacc->offset - sad->left_offset,
3032 lacc, sad->new_gsi, true);
3033 else
3034 rhs = build_ref_for_model (sad->loc, sad->assignment_rhs,
3035 lacc->offset - sad->left_offset,
3036 lacc, sad->new_gsi, true);
3037 if (lacc->grp_partial_lhs)
3038 rhs = force_gimple_operand_gsi (sad->new_gsi,
3039 rhs, true, NULL_TREE,
3040 false, GSI_NEW_STMT);
3041 }
3042
3043 stmt = gimple_build_assign (get_access_replacement (lacc), rhs);
3044 gsi_insert_after (sad->new_gsi, stmt, GSI_NEW_STMT);
3045 gimple_set_location (stmt, sad->loc);
3046 update_stmt (stmt);
3047 sra_stats.subreplacements++;
3048 }
3049 else
3050 {
3051 if (sad->refreshed == SRA_UDH_NONE
3052 && lacc->grp_read && !lacc->grp_covered)
3053 handle_unscalarized_data_in_subtree (sad);
3054
3055 if (lacc && lacc->grp_to_be_debug_replaced)
3056 {
3057 gdebug *ds;
3058 tree drhs;
3059 struct access *racc = find_access_in_subtree (sad->top_racc,
3060 offset,
3061 lacc->size);
3062
3063 if (racc && racc->grp_to_be_replaced)
3064 {
3065 if (racc->grp_write)
3066 drhs = get_access_replacement (racc);
3067 else
3068 drhs = NULL;
3069 }
3070 else if (sad->refreshed == SRA_UDH_LEFT)
3071 drhs = build_debug_ref_for_model (sad->loc, lacc->base,
3072 lacc->offset, lacc);
3073 else if (sad->refreshed == SRA_UDH_RIGHT)
3074 drhs = build_debug_ref_for_model (sad->loc, sad->top_racc->base,
3075 offset, lacc);
3076 else
3077 drhs = NULL_TREE;
3078 if (drhs
3079 && !useless_type_conversion_p (lacc->type, TREE_TYPE (drhs)))
3080 drhs = fold_build1_loc (sad->loc, VIEW_CONVERT_EXPR,
3081 lacc->type, drhs);
3082 ds = gimple_build_debug_bind (get_access_replacement (lacc),
3083 drhs, gsi_stmt (sad->old_gsi));
3084 gsi_insert_after (sad->new_gsi, ds, GSI_NEW_STMT);
3085 }
3086 }
3087
3088 if (lacc->first_child)
3089 load_assign_lhs_subreplacements (lacc, sad);
3090 }
3091 }
3092
3093 /* Result code for SRA assignment modification. */
3094 enum assignment_mod_result { SRA_AM_NONE, /* nothing done for the stmt */
3095 SRA_AM_MODIFIED, /* stmt changed but not
3096 removed */
3097 SRA_AM_REMOVED }; /* stmt eliminated */
3098
3099 /* Modify assignments with a CONSTRUCTOR on their RHS. STMT contains a pointer
3100 to the assignment and GSI is the statement iterator pointing at it. Returns
3101 the same values as sra_modify_assign. */
3102
3103 static enum assignment_mod_result
3104 sra_modify_constructor_assign (gimple stmt, gimple_stmt_iterator *gsi)
3105 {
3106 tree lhs = gimple_assign_lhs (stmt);
3107 struct access *acc = get_access_for_expr (lhs);
3108 if (!acc)
3109 return SRA_AM_NONE;
3110 location_t loc = gimple_location (stmt);
3111
3112 if (gimple_clobber_p (stmt))
3113 {
3114 /* Clobber the replacement variable. */
3115 clobber_subtree (acc, gsi, !acc->grp_covered, loc);
3116 /* Remove clobbers of fully scalarized variables, they are dead. */
3117 if (acc->grp_covered)
3118 {
3119 unlink_stmt_vdef (stmt);
3120 gsi_remove (gsi, true);
3121 release_defs (stmt);
3122 return SRA_AM_REMOVED;
3123 }
3124 else
3125 return SRA_AM_MODIFIED;
3126 }
3127
3128 if (vec_safe_length (CONSTRUCTOR_ELTS (gimple_assign_rhs1 (stmt))) > 0)
3129 {
3130 /* I have never seen this code path trigger but if it can happen the
3131 following should handle it gracefully. */
3132 if (access_has_children_p (acc))
3133 generate_subtree_copies (acc->first_child, lhs, acc->offset, 0, 0, gsi,
3134 true, true, loc);
3135 return SRA_AM_MODIFIED;
3136 }
3137
3138 if (acc->grp_covered)
3139 {
3140 init_subtree_with_zero (acc, gsi, false, loc);
3141 unlink_stmt_vdef (stmt);
3142 gsi_remove (gsi, true);
3143 release_defs (stmt);
3144 return SRA_AM_REMOVED;
3145 }
3146 else
3147 {
3148 init_subtree_with_zero (acc, gsi, true, loc);
3149 return SRA_AM_MODIFIED;
3150 }
3151 }
3152
3153 /* Create and return a new suitable default definition SSA_NAME for RACC which
3154 is an access describing an uninitialized part of an aggregate that is being
3155 loaded. */
3156
3157 static tree
3158 get_repl_default_def_ssa_name (struct access *racc)
3159 {
3160 gcc_checking_assert (!racc->grp_to_be_replaced
3161 && !racc->grp_to_be_debug_replaced);
3162 if (!racc->replacement_decl)
3163 racc->replacement_decl = create_access_replacement (racc);
3164 return get_or_create_ssa_default_def (cfun, racc->replacement_decl);
3165 }
3166
3167 /* Return true if REF has an VIEW_CONVERT_EXPR or a COMPONENT_REF with a
3168 bit-field field declaration somewhere in it. */
3169
3170 static inline bool
3171 contains_vce_or_bfcref_p (const_tree ref)
3172 {
3173 while (handled_component_p (ref))
3174 {
3175 if (TREE_CODE (ref) == VIEW_CONVERT_EXPR
3176 || (TREE_CODE (ref) == COMPONENT_REF
3177 && DECL_BIT_FIELD (TREE_OPERAND (ref, 1))))
3178 return true;
3179 ref = TREE_OPERAND (ref, 0);
3180 }
3181
3182 return false;
3183 }
3184
3185 /* Examine both sides of the assignment statement pointed to by STMT, replace
3186 them with a scalare replacement if there is one and generate copying of
3187 replacements if scalarized aggregates have been used in the assignment. GSI
3188 is used to hold generated statements for type conversions and subtree
3189 copying. */
3190
3191 static enum assignment_mod_result
3192 sra_modify_assign (gimple stmt, gimple_stmt_iterator *gsi)
3193 {
3194 struct access *lacc, *racc;
3195 tree lhs, rhs;
3196 bool modify_this_stmt = false;
3197 bool force_gimple_rhs = false;
3198 location_t loc;
3199 gimple_stmt_iterator orig_gsi = *gsi;
3200
3201 if (!gimple_assign_single_p (stmt))
3202 return SRA_AM_NONE;
3203 lhs = gimple_assign_lhs (stmt);
3204 rhs = gimple_assign_rhs1 (stmt);
3205
3206 if (TREE_CODE (rhs) == CONSTRUCTOR)
3207 return sra_modify_constructor_assign (stmt, gsi);
3208
3209 if (TREE_CODE (rhs) == REALPART_EXPR || TREE_CODE (lhs) == REALPART_EXPR
3210 || TREE_CODE (rhs) == IMAGPART_EXPR || TREE_CODE (lhs) == IMAGPART_EXPR
3211 || TREE_CODE (rhs) == BIT_FIELD_REF || TREE_CODE (lhs) == BIT_FIELD_REF)
3212 {
3213 modify_this_stmt = sra_modify_expr (gimple_assign_rhs1_ptr (stmt),
3214 gsi, false);
3215 modify_this_stmt |= sra_modify_expr (gimple_assign_lhs_ptr (stmt),
3216 gsi, true);
3217 return modify_this_stmt ? SRA_AM_MODIFIED : SRA_AM_NONE;
3218 }
3219
3220 lacc = get_access_for_expr (lhs);
3221 racc = get_access_for_expr (rhs);
3222 if (!lacc && !racc)
3223 return SRA_AM_NONE;
3224
3225 loc = gimple_location (stmt);
3226 if (lacc && lacc->grp_to_be_replaced)
3227 {
3228 lhs = get_access_replacement (lacc);
3229 gimple_assign_set_lhs (stmt, lhs);
3230 modify_this_stmt = true;
3231 if (lacc->grp_partial_lhs)
3232 force_gimple_rhs = true;
3233 sra_stats.exprs++;
3234 }
3235
3236 if (racc && racc->grp_to_be_replaced)
3237 {
3238 rhs = get_access_replacement (racc);
3239 modify_this_stmt = true;
3240 if (racc->grp_partial_lhs)
3241 force_gimple_rhs = true;
3242 sra_stats.exprs++;
3243 }
3244 else if (racc
3245 && !racc->grp_unscalarized_data
3246 && TREE_CODE (lhs) == SSA_NAME
3247 && !access_has_replacements_p (racc))
3248 {
3249 rhs = get_repl_default_def_ssa_name (racc);
3250 modify_this_stmt = true;
3251 sra_stats.exprs++;
3252 }
3253
3254 if (modify_this_stmt)
3255 {
3256 if (!useless_type_conversion_p (TREE_TYPE (lhs), TREE_TYPE (rhs)))
3257 {
3258 /* If we can avoid creating a VIEW_CONVERT_EXPR do so.
3259 ??? This should move to fold_stmt which we simply should
3260 call after building a VIEW_CONVERT_EXPR here. */
3261 if (AGGREGATE_TYPE_P (TREE_TYPE (lhs))
3262 && !contains_bitfld_component_ref_p (lhs))
3263 {
3264 lhs = build_ref_for_model (loc, lhs, 0, racc, gsi, false);
3265 gimple_assign_set_lhs (stmt, lhs);
3266 }
3267 else if (AGGREGATE_TYPE_P (TREE_TYPE (rhs))
3268 && !contains_vce_or_bfcref_p (rhs))
3269 rhs = build_ref_for_model (loc, rhs, 0, lacc, gsi, false);
3270
3271 if (!useless_type_conversion_p (TREE_TYPE (lhs), TREE_TYPE (rhs)))
3272 {
3273 rhs = fold_build1_loc (loc, VIEW_CONVERT_EXPR, TREE_TYPE (lhs),
3274 rhs);
3275 if (is_gimple_reg_type (TREE_TYPE (lhs))
3276 && TREE_CODE (lhs) != SSA_NAME)
3277 force_gimple_rhs = true;
3278 }
3279 }
3280 }
3281
3282 if (lacc && lacc->grp_to_be_debug_replaced)
3283 {
3284 tree dlhs = get_access_replacement (lacc);
3285 tree drhs = unshare_expr (rhs);
3286 if (!useless_type_conversion_p (TREE_TYPE (dlhs), TREE_TYPE (drhs)))
3287 {
3288 if (AGGREGATE_TYPE_P (TREE_TYPE (drhs))
3289 && !contains_vce_or_bfcref_p (drhs))
3290 drhs = build_debug_ref_for_model (loc, drhs, 0, lacc);
3291 if (drhs
3292 && !useless_type_conversion_p (TREE_TYPE (dlhs),
3293 TREE_TYPE (drhs)))
3294 drhs = fold_build1_loc (loc, VIEW_CONVERT_EXPR,
3295 TREE_TYPE (dlhs), drhs);
3296 }
3297 gdebug *ds = gimple_build_debug_bind (dlhs, drhs, stmt);
3298 gsi_insert_before (gsi, ds, GSI_SAME_STMT);
3299 }
3300
3301 /* From this point on, the function deals with assignments in between
3302 aggregates when at least one has scalar reductions of some of its
3303 components. There are three possible scenarios: Both the LHS and RHS have
3304 to-be-scalarized components, 2) only the RHS has or 3) only the LHS has.
3305
3306 In the first case, we would like to load the LHS components from RHS
3307 components whenever possible. If that is not possible, we would like to
3308 read it directly from the RHS (after updating it by storing in it its own
3309 components). If there are some necessary unscalarized data in the LHS,
3310 those will be loaded by the original assignment too. If neither of these
3311 cases happen, the original statement can be removed. Most of this is done
3312 by load_assign_lhs_subreplacements.
3313
3314 In the second case, we would like to store all RHS scalarized components
3315 directly into LHS and if they cover the aggregate completely, remove the
3316 statement too. In the third case, we want the LHS components to be loaded
3317 directly from the RHS (DSE will remove the original statement if it
3318 becomes redundant).
3319
3320 This is a bit complex but manageable when types match and when unions do
3321 not cause confusion in a way that we cannot really load a component of LHS
3322 from the RHS or vice versa (the access representing this level can have
3323 subaccesses that are accessible only through a different union field at a
3324 higher level - different from the one used in the examined expression).
3325 Unions are fun.
3326
3327 Therefore, I specially handle a fourth case, happening when there is a
3328 specific type cast or it is impossible to locate a scalarized subaccess on
3329 the other side of the expression. If that happens, I simply "refresh" the
3330 RHS by storing in it is scalarized components leave the original statement
3331 there to do the copying and then load the scalar replacements of the LHS.
3332 This is what the first branch does. */
3333
3334 if (modify_this_stmt
3335 || gimple_has_volatile_ops (stmt)
3336 || contains_vce_or_bfcref_p (rhs)
3337 || contains_vce_or_bfcref_p (lhs)
3338 || stmt_ends_bb_p (stmt))
3339 {
3340 if (access_has_children_p (racc))
3341 generate_subtree_copies (racc->first_child, rhs, racc->offset, 0, 0,
3342 gsi, false, false, loc);
3343 if (access_has_children_p (lacc))
3344 {
3345 gimple_stmt_iterator alt_gsi = gsi_none ();
3346 if (stmt_ends_bb_p (stmt))
3347 {
3348 alt_gsi = gsi_start_edge (single_non_eh_succ (gsi_bb (*gsi)));
3349 gsi = &alt_gsi;
3350 }
3351 generate_subtree_copies (lacc->first_child, lhs, lacc->offset, 0, 0,
3352 gsi, true, true, loc);
3353 }
3354 sra_stats.separate_lhs_rhs_handling++;
3355
3356 /* This gimplification must be done after generate_subtree_copies,
3357 lest we insert the subtree copies in the middle of the gimplified
3358 sequence. */
3359 if (force_gimple_rhs)
3360 rhs = force_gimple_operand_gsi (&orig_gsi, rhs, true, NULL_TREE,
3361 true, GSI_SAME_STMT);
3362 if (gimple_assign_rhs1 (stmt) != rhs)
3363 {
3364 modify_this_stmt = true;
3365 gimple_assign_set_rhs_from_tree (&orig_gsi, rhs);
3366 gcc_assert (stmt == gsi_stmt (orig_gsi));
3367 }
3368
3369 return modify_this_stmt ? SRA_AM_MODIFIED : SRA_AM_NONE;
3370 }
3371 else
3372 {
3373 if (access_has_children_p (lacc)
3374 && access_has_children_p (racc)
3375 /* When an access represents an unscalarizable region, it usually
3376 represents accesses with variable offset and thus must not be used
3377 to generate new memory accesses. */
3378 && !lacc->grp_unscalarizable_region
3379 && !racc->grp_unscalarizable_region)
3380 {
3381 struct subreplacement_assignment_data sad;
3382
3383 sad.left_offset = lacc->offset;
3384 sad.assignment_lhs = lhs;
3385 sad.assignment_rhs = rhs;
3386 sad.top_racc = racc;
3387 sad.old_gsi = *gsi;
3388 sad.new_gsi = gsi;
3389 sad.loc = gimple_location (stmt);
3390 sad.refreshed = SRA_UDH_NONE;
3391
3392 if (lacc->grp_read && !lacc->grp_covered)
3393 handle_unscalarized_data_in_subtree (&sad);
3394
3395 load_assign_lhs_subreplacements (lacc, &sad);
3396 if (sad.refreshed != SRA_UDH_RIGHT)
3397 {
3398 gsi_next (gsi);
3399 unlink_stmt_vdef (stmt);
3400 gsi_remove (&sad.old_gsi, true);
3401 release_defs (stmt);
3402 sra_stats.deleted++;
3403 return SRA_AM_REMOVED;
3404 }
3405 }
3406 else
3407 {
3408 if (access_has_children_p (racc)
3409 && !racc->grp_unscalarized_data)
3410 {
3411 if (dump_file)
3412 {
3413 fprintf (dump_file, "Removing load: ");
3414 print_gimple_stmt (dump_file, stmt, 0, 0);
3415 }
3416 generate_subtree_copies (racc->first_child, lhs,
3417 racc->offset, 0, 0, gsi,
3418 false, false, loc);
3419 gcc_assert (stmt == gsi_stmt (*gsi));
3420 unlink_stmt_vdef (stmt);
3421 gsi_remove (gsi, true);
3422 release_defs (stmt);
3423 sra_stats.deleted++;
3424 return SRA_AM_REMOVED;
3425 }
3426 /* Restore the aggregate RHS from its components so the
3427 prevailing aggregate copy does the right thing. */
3428 if (access_has_children_p (racc))
3429 generate_subtree_copies (racc->first_child, rhs, racc->offset, 0, 0,
3430 gsi, false, false, loc);
3431 /* Re-load the components of the aggregate copy destination.
3432 But use the RHS aggregate to load from to expose more
3433 optimization opportunities. */
3434 if (access_has_children_p (lacc))
3435 generate_subtree_copies (lacc->first_child, rhs, lacc->offset,
3436 0, 0, gsi, true, true, loc);
3437 }
3438
3439 return SRA_AM_NONE;
3440 }
3441 }
3442
3443 /* Traverse the function body and all modifications as decided in
3444 analyze_all_variable_accesses. Return true iff the CFG has been
3445 changed. */
3446
3447 static bool
3448 sra_modify_function_body (void)
3449 {
3450 bool cfg_changed = false;
3451 basic_block bb;
3452
3453 FOR_EACH_BB_FN (bb, cfun)
3454 {
3455 gimple_stmt_iterator gsi = gsi_start_bb (bb);
3456 while (!gsi_end_p (gsi))
3457 {
3458 gimple stmt = gsi_stmt (gsi);
3459 enum assignment_mod_result assign_result;
3460 bool modified = false, deleted = false;
3461 tree *t;
3462 unsigned i;
3463
3464 switch (gimple_code (stmt))
3465 {
3466 case GIMPLE_RETURN:
3467 t = gimple_return_retval_ptr (as_a <greturn *> (stmt));
3468 if (*t != NULL_TREE)
3469 modified |= sra_modify_expr (t, &gsi, false);
3470 break;
3471
3472 case GIMPLE_ASSIGN:
3473 assign_result = sra_modify_assign (stmt, &gsi);
3474 modified |= assign_result == SRA_AM_MODIFIED;
3475 deleted = assign_result == SRA_AM_REMOVED;
3476 break;
3477
3478 case GIMPLE_CALL:
3479 /* Operands must be processed before the lhs. */
3480 for (i = 0; i < gimple_call_num_args (stmt); i++)
3481 {
3482 t = gimple_call_arg_ptr (stmt, i);
3483 modified |= sra_modify_expr (t, &gsi, false);
3484 }
3485
3486 if (gimple_call_lhs (stmt))
3487 {
3488 t = gimple_call_lhs_ptr (stmt);
3489 modified |= sra_modify_expr (t, &gsi, true);
3490 }
3491 break;
3492
3493 case GIMPLE_ASM:
3494 {
3495 gasm *asm_stmt = as_a <gasm *> (stmt);
3496 for (i = 0; i < gimple_asm_ninputs (asm_stmt); i++)
3497 {
3498 t = &TREE_VALUE (gimple_asm_input_op (asm_stmt, i));
3499 modified |= sra_modify_expr (t, &gsi, false);
3500 }
3501 for (i = 0; i < gimple_asm_noutputs (asm_stmt); i++)
3502 {
3503 t = &TREE_VALUE (gimple_asm_output_op (asm_stmt, i));
3504 modified |= sra_modify_expr (t, &gsi, true);
3505 }
3506 }
3507 break;
3508
3509 default:
3510 break;
3511 }
3512
3513 if (modified)
3514 {
3515 update_stmt (stmt);
3516 if (maybe_clean_eh_stmt (stmt)
3517 && gimple_purge_dead_eh_edges (gimple_bb (stmt)))
3518 cfg_changed = true;
3519 }
3520 if (!deleted)
3521 gsi_next (&gsi);
3522 }
3523 }
3524
3525 gsi_commit_edge_inserts ();
3526 return cfg_changed;
3527 }
3528
3529 /* Generate statements initializing scalar replacements of parts of function
3530 parameters. */
3531
3532 static void
3533 initialize_parameter_reductions (void)
3534 {
3535 gimple_stmt_iterator gsi;
3536 gimple_seq seq = NULL;
3537 tree parm;
3538
3539 gsi = gsi_start (seq);
3540 for (parm = DECL_ARGUMENTS (current_function_decl);
3541 parm;
3542 parm = DECL_CHAIN (parm))
3543 {
3544 vec<access_p> *access_vec;
3545 struct access *access;
3546
3547 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
3548 continue;
3549 access_vec = get_base_access_vector (parm);
3550 if (!access_vec)
3551 continue;
3552
3553 for (access = (*access_vec)[0];
3554 access;
3555 access = access->next_grp)
3556 generate_subtree_copies (access, parm, 0, 0, 0, &gsi, true, true,
3557 EXPR_LOCATION (parm));
3558 }
3559
3560 seq = gsi_seq (gsi);
3561 if (seq)
3562 gsi_insert_seq_on_edge_immediate (single_succ_edge (ENTRY_BLOCK_PTR_FOR_FN (cfun)), seq);
3563 }
3564
3565 /* The "main" function of intraprocedural SRA passes. Runs the analysis and if
3566 it reveals there are components of some aggregates to be scalarized, it runs
3567 the required transformations. */
3568 static unsigned int
3569 perform_intra_sra (void)
3570 {
3571 int ret = 0;
3572 sra_initialize ();
3573
3574 if (!find_var_candidates ())
3575 goto out;
3576
3577 if (!scan_function ())
3578 goto out;
3579
3580 if (!analyze_all_variable_accesses ())
3581 goto out;
3582
3583 if (sra_modify_function_body ())
3584 ret = TODO_update_ssa | TODO_cleanup_cfg;
3585 else
3586 ret = TODO_update_ssa;
3587 initialize_parameter_reductions ();
3588
3589 statistics_counter_event (cfun, "Scalar replacements created",
3590 sra_stats.replacements);
3591 statistics_counter_event (cfun, "Modified expressions", sra_stats.exprs);
3592 statistics_counter_event (cfun, "Subtree copy stmts",
3593 sra_stats.subtree_copies);
3594 statistics_counter_event (cfun, "Subreplacement stmts",
3595 sra_stats.subreplacements);
3596 statistics_counter_event (cfun, "Deleted stmts", sra_stats.deleted);
3597 statistics_counter_event (cfun, "Separate LHS and RHS handling",
3598 sra_stats.separate_lhs_rhs_handling);
3599
3600 out:
3601 sra_deinitialize ();
3602 return ret;
3603 }
3604
3605 /* Perform early intraprocedural SRA. */
3606 static unsigned int
3607 early_intra_sra (void)
3608 {
3609 sra_mode = SRA_MODE_EARLY_INTRA;
3610 return perform_intra_sra ();
3611 }
3612
3613 /* Perform "late" intraprocedural SRA. */
3614 static unsigned int
3615 late_intra_sra (void)
3616 {
3617 sra_mode = SRA_MODE_INTRA;
3618 return perform_intra_sra ();
3619 }
3620
3621
3622 static bool
3623 gate_intra_sra (void)
3624 {
3625 return flag_tree_sra != 0 && dbg_cnt (tree_sra);
3626 }
3627
3628
3629 namespace {
3630
3631 const pass_data pass_data_sra_early =
3632 {
3633 GIMPLE_PASS, /* type */
3634 "esra", /* name */
3635 OPTGROUP_NONE, /* optinfo_flags */
3636 TV_TREE_SRA, /* tv_id */
3637 ( PROP_cfg | PROP_ssa ), /* properties_required */
3638 0, /* properties_provided */
3639 0, /* properties_destroyed */
3640 0, /* todo_flags_start */
3641 TODO_update_ssa, /* todo_flags_finish */
3642 };
3643
3644 class pass_sra_early : public gimple_opt_pass
3645 {
3646 public:
3647 pass_sra_early (gcc::context *ctxt)
3648 : gimple_opt_pass (pass_data_sra_early, ctxt)
3649 {}
3650
3651 /* opt_pass methods: */
3652 virtual bool gate (function *) { return gate_intra_sra (); }
3653 virtual unsigned int execute (function *) { return early_intra_sra (); }
3654
3655 }; // class pass_sra_early
3656
3657 } // anon namespace
3658
3659 gimple_opt_pass *
3660 make_pass_sra_early (gcc::context *ctxt)
3661 {
3662 return new pass_sra_early (ctxt);
3663 }
3664
3665 namespace {
3666
3667 const pass_data pass_data_sra =
3668 {
3669 GIMPLE_PASS, /* type */
3670 "sra", /* name */
3671 OPTGROUP_NONE, /* optinfo_flags */
3672 TV_TREE_SRA, /* tv_id */
3673 ( PROP_cfg | PROP_ssa ), /* properties_required */
3674 0, /* properties_provided */
3675 0, /* properties_destroyed */
3676 TODO_update_address_taken, /* todo_flags_start */
3677 TODO_update_ssa, /* todo_flags_finish */
3678 };
3679
3680 class pass_sra : public gimple_opt_pass
3681 {
3682 public:
3683 pass_sra (gcc::context *ctxt)
3684 : gimple_opt_pass (pass_data_sra, ctxt)
3685 {}
3686
3687 /* opt_pass methods: */
3688 virtual bool gate (function *) { return gate_intra_sra (); }
3689 virtual unsigned int execute (function *) { return late_intra_sra (); }
3690
3691 }; // class pass_sra
3692
3693 } // anon namespace
3694
3695 gimple_opt_pass *
3696 make_pass_sra (gcc::context *ctxt)
3697 {
3698 return new pass_sra (ctxt);
3699 }
3700
3701
3702 /* Return true iff PARM (which must be a parm_decl) is an unused scalar
3703 parameter. */
3704
3705 static bool
3706 is_unused_scalar_param (tree parm)
3707 {
3708 tree name;
3709 return (is_gimple_reg (parm)
3710 && (!(name = ssa_default_def (cfun, parm))
3711 || has_zero_uses (name)));
3712 }
3713
3714 /* Scan immediate uses of a default definition SSA name of a parameter PARM and
3715 examine whether there are any direct or otherwise infeasible ones. If so,
3716 return true, otherwise return false. PARM must be a gimple register with a
3717 non-NULL default definition. */
3718
3719 static bool
3720 ptr_parm_has_direct_uses (tree parm)
3721 {
3722 imm_use_iterator ui;
3723 gimple stmt;
3724 tree name = ssa_default_def (cfun, parm);
3725 bool ret = false;
3726
3727 FOR_EACH_IMM_USE_STMT (stmt, ui, name)
3728 {
3729 int uses_ok = 0;
3730 use_operand_p use_p;
3731
3732 if (is_gimple_debug (stmt))
3733 continue;
3734
3735 /* Valid uses include dereferences on the lhs and the rhs. */
3736 if (gimple_has_lhs (stmt))
3737 {
3738 tree lhs = gimple_get_lhs (stmt);
3739 while (handled_component_p (lhs))
3740 lhs = TREE_OPERAND (lhs, 0);
3741 if (TREE_CODE (lhs) == MEM_REF
3742 && TREE_OPERAND (lhs, 0) == name
3743 && integer_zerop (TREE_OPERAND (lhs, 1))
3744 && types_compatible_p (TREE_TYPE (lhs),
3745 TREE_TYPE (TREE_TYPE (name)))
3746 && !TREE_THIS_VOLATILE (lhs))
3747 uses_ok++;
3748 }
3749 if (gimple_assign_single_p (stmt))
3750 {
3751 tree rhs = gimple_assign_rhs1 (stmt);
3752 while (handled_component_p (rhs))
3753 rhs = TREE_OPERAND (rhs, 0);
3754 if (TREE_CODE (rhs) == MEM_REF
3755 && TREE_OPERAND (rhs, 0) == name
3756 && integer_zerop (TREE_OPERAND (rhs, 1))
3757 && types_compatible_p (TREE_TYPE (rhs),
3758 TREE_TYPE (TREE_TYPE (name)))
3759 && !TREE_THIS_VOLATILE (rhs))
3760 uses_ok++;
3761 }
3762 else if (is_gimple_call (stmt))
3763 {
3764 unsigned i;
3765 for (i = 0; i < gimple_call_num_args (stmt); ++i)
3766 {
3767 tree arg = gimple_call_arg (stmt, i);
3768 while (handled_component_p (arg))
3769 arg = TREE_OPERAND (arg, 0);
3770 if (TREE_CODE (arg) == MEM_REF
3771 && TREE_OPERAND (arg, 0) == name
3772 && integer_zerop (TREE_OPERAND (arg, 1))
3773 && types_compatible_p (TREE_TYPE (arg),
3774 TREE_TYPE (TREE_TYPE (name)))
3775 && !TREE_THIS_VOLATILE (arg))
3776 uses_ok++;
3777 }
3778 }
3779
3780 /* If the number of valid uses does not match the number of
3781 uses in this stmt there is an unhandled use. */
3782 FOR_EACH_IMM_USE_ON_STMT (use_p, ui)
3783 --uses_ok;
3784
3785 if (uses_ok != 0)
3786 ret = true;
3787
3788 if (ret)
3789 BREAK_FROM_IMM_USE_STMT (ui);
3790 }
3791
3792 return ret;
3793 }
3794
3795 /* Identify candidates for reduction for IPA-SRA based on their type and mark
3796 them in candidate_bitmap. Note that these do not necessarily include
3797 parameter which are unused and thus can be removed. Return true iff any
3798 such candidate has been found. */
3799
3800 static bool
3801 find_param_candidates (void)
3802 {
3803 tree parm;
3804 int count = 0;
3805 bool ret = false;
3806 const char *msg;
3807
3808 for (parm = DECL_ARGUMENTS (current_function_decl);
3809 parm;
3810 parm = DECL_CHAIN (parm))
3811 {
3812 tree type = TREE_TYPE (parm);
3813 tree_node **slot;
3814
3815 count++;
3816
3817 if (TREE_THIS_VOLATILE (parm)
3818 || TREE_ADDRESSABLE (parm)
3819 || (!is_gimple_reg_type (type) && is_va_list_type (type)))
3820 continue;
3821
3822 if (is_unused_scalar_param (parm))
3823 {
3824 ret = true;
3825 continue;
3826 }
3827
3828 if (POINTER_TYPE_P (type))
3829 {
3830 type = TREE_TYPE (type);
3831
3832 if (TREE_CODE (type) == FUNCTION_TYPE
3833 || TYPE_VOLATILE (type)
3834 || (TREE_CODE (type) == ARRAY_TYPE
3835 && TYPE_NONALIASED_COMPONENT (type))
3836 || !is_gimple_reg (parm)
3837 || is_va_list_type (type)
3838 || ptr_parm_has_direct_uses (parm))
3839 continue;
3840 }
3841 else if (!AGGREGATE_TYPE_P (type))
3842 continue;
3843
3844 if (!COMPLETE_TYPE_P (type)
3845 || !tree_fits_uhwi_p (TYPE_SIZE (type))
3846 || tree_to_uhwi (TYPE_SIZE (type)) == 0
3847 || (AGGREGATE_TYPE_P (type)
3848 && type_internals_preclude_sra_p (type, &msg)))
3849 continue;
3850
3851 bitmap_set_bit (candidate_bitmap, DECL_UID (parm));
3852 slot = candidates->find_slot_with_hash (parm, DECL_UID (parm), INSERT);
3853 *slot = parm;
3854
3855 ret = true;
3856 if (dump_file && (dump_flags & TDF_DETAILS))
3857 {
3858 fprintf (dump_file, "Candidate (%d): ", DECL_UID (parm));
3859 print_generic_expr (dump_file, parm, 0);
3860 fprintf (dump_file, "\n");
3861 }
3862 }
3863
3864 func_param_count = count;
3865 return ret;
3866 }
3867
3868 /* Callback of walk_aliased_vdefs, marks the access passed as DATA as
3869 maybe_modified. */
3870
3871 static bool
3872 mark_maybe_modified (ao_ref *ao ATTRIBUTE_UNUSED, tree vdef ATTRIBUTE_UNUSED,
3873 void *data)
3874 {
3875 struct access *repr = (struct access *) data;
3876
3877 repr->grp_maybe_modified = 1;
3878 return true;
3879 }
3880
3881 /* Analyze what representatives (in linked lists accessible from
3882 REPRESENTATIVES) can be modified by side effects of statements in the
3883 current function. */
3884
3885 static void
3886 analyze_modified_params (vec<access_p> representatives)
3887 {
3888 int i;
3889
3890 for (i = 0; i < func_param_count; i++)
3891 {
3892 struct access *repr;
3893
3894 for (repr = representatives[i];
3895 repr;
3896 repr = repr->next_grp)
3897 {
3898 struct access *access;
3899 bitmap visited;
3900 ao_ref ar;
3901
3902 if (no_accesses_p (repr))
3903 continue;
3904 if (!POINTER_TYPE_P (TREE_TYPE (repr->base))
3905 || repr->grp_maybe_modified)
3906 continue;
3907
3908 ao_ref_init (&ar, repr->expr);
3909 visited = BITMAP_ALLOC (NULL);
3910 for (access = repr; access; access = access->next_sibling)
3911 {
3912 /* All accesses are read ones, otherwise grp_maybe_modified would
3913 be trivially set. */
3914 walk_aliased_vdefs (&ar, gimple_vuse (access->stmt),
3915 mark_maybe_modified, repr, &visited);
3916 if (repr->grp_maybe_modified)
3917 break;
3918 }
3919 BITMAP_FREE (visited);
3920 }
3921 }
3922 }
3923
3924 /* Propagate distances in bb_dereferences in the opposite direction than the
3925 control flow edges, in each step storing the maximum of the current value
3926 and the minimum of all successors. These steps are repeated until the table
3927 stabilizes. Note that BBs which might terminate the functions (according to
3928 final_bbs bitmap) never updated in this way. */
3929
3930 static void
3931 propagate_dereference_distances (void)
3932 {
3933 basic_block bb;
3934
3935 auto_vec<basic_block> queue (last_basic_block_for_fn (cfun));
3936 queue.quick_push (ENTRY_BLOCK_PTR_FOR_FN (cfun));
3937 FOR_EACH_BB_FN (bb, cfun)
3938 {
3939 queue.quick_push (bb);
3940 bb->aux = bb;
3941 }
3942
3943 while (!queue.is_empty ())
3944 {
3945 edge_iterator ei;
3946 edge e;
3947 bool change = false;
3948 int i;
3949
3950 bb = queue.pop ();
3951 bb->aux = NULL;
3952
3953 if (bitmap_bit_p (final_bbs, bb->index))
3954 continue;
3955
3956 for (i = 0; i < func_param_count; i++)
3957 {
3958 int idx = bb->index * func_param_count + i;
3959 bool first = true;
3960 HOST_WIDE_INT inh = 0;
3961
3962 FOR_EACH_EDGE (e, ei, bb->succs)
3963 {
3964 int succ_idx = e->dest->index * func_param_count + i;
3965
3966 if (e->src == EXIT_BLOCK_PTR_FOR_FN (cfun))
3967 continue;
3968
3969 if (first)
3970 {
3971 first = false;
3972 inh = bb_dereferences [succ_idx];
3973 }
3974 else if (bb_dereferences [succ_idx] < inh)
3975 inh = bb_dereferences [succ_idx];
3976 }
3977
3978 if (!first && bb_dereferences[idx] < inh)
3979 {
3980 bb_dereferences[idx] = inh;
3981 change = true;
3982 }
3983 }
3984
3985 if (change && !bitmap_bit_p (final_bbs, bb->index))
3986 FOR_EACH_EDGE (e, ei, bb->preds)
3987 {
3988 if (e->src->aux)
3989 continue;
3990
3991 e->src->aux = e->src;
3992 queue.quick_push (e->src);
3993 }
3994 }
3995 }
3996
3997 /* Dump a dereferences TABLE with heading STR to file F. */
3998
3999 static void
4000 dump_dereferences_table (FILE *f, const char *str, HOST_WIDE_INT *table)
4001 {
4002 basic_block bb;
4003
4004 fprintf (dump_file, "%s", str);
4005 FOR_BB_BETWEEN (bb, ENTRY_BLOCK_PTR_FOR_FN (cfun),
4006 EXIT_BLOCK_PTR_FOR_FN (cfun), next_bb)
4007 {
4008 fprintf (f, "%4i %i ", bb->index, bitmap_bit_p (final_bbs, bb->index));
4009 if (bb != EXIT_BLOCK_PTR_FOR_FN (cfun))
4010 {
4011 int i;
4012 for (i = 0; i < func_param_count; i++)
4013 {
4014 int idx = bb->index * func_param_count + i;
4015 fprintf (f, " %4" HOST_WIDE_INT_PRINT "d", table[idx]);
4016 }
4017 }
4018 fprintf (f, "\n");
4019 }
4020 fprintf (dump_file, "\n");
4021 }
4022
4023 /* Determine what (parts of) parameters passed by reference that are not
4024 assigned to are not certainly dereferenced in this function and thus the
4025 dereferencing cannot be safely moved to the caller without potentially
4026 introducing a segfault. Mark such REPRESENTATIVES as
4027 grp_not_necessarilly_dereferenced.
4028
4029 The dereferenced maximum "distance," i.e. the offset + size of the accessed
4030 part is calculated rather than simple booleans are calculated for each
4031 pointer parameter to handle cases when only a fraction of the whole
4032 aggregate is allocated (see testsuite/gcc.c-torture/execute/ipa-sra-2.c for
4033 an example).
4034
4035 The maximum dereference distances for each pointer parameter and BB are
4036 already stored in bb_dereference. This routine simply propagates these
4037 values upwards by propagate_dereference_distances and then compares the
4038 distances of individual parameters in the ENTRY BB to the equivalent
4039 distances of each representative of a (fraction of a) parameter. */
4040
4041 static void
4042 analyze_caller_dereference_legality (vec<access_p> representatives)
4043 {
4044 int i;
4045
4046 if (dump_file && (dump_flags & TDF_DETAILS))
4047 dump_dereferences_table (dump_file,
4048 "Dereference table before propagation:\n",
4049 bb_dereferences);
4050
4051 propagate_dereference_distances ();
4052
4053 if (dump_file && (dump_flags & TDF_DETAILS))
4054 dump_dereferences_table (dump_file,
4055 "Dereference table after propagation:\n",
4056 bb_dereferences);
4057
4058 for (i = 0; i < func_param_count; i++)
4059 {
4060 struct access *repr = representatives[i];
4061 int idx = ENTRY_BLOCK_PTR_FOR_FN (cfun)->index * func_param_count + i;
4062
4063 if (!repr || no_accesses_p (repr))
4064 continue;
4065
4066 do
4067 {
4068 if ((repr->offset + repr->size) > bb_dereferences[idx])
4069 repr->grp_not_necessarilly_dereferenced = 1;
4070 repr = repr->next_grp;
4071 }
4072 while (repr);
4073 }
4074 }
4075
4076 /* Return the representative access for the parameter declaration PARM if it is
4077 a scalar passed by reference which is not written to and the pointer value
4078 is not used directly. Thus, if it is legal to dereference it in the caller
4079 and we can rule out modifications through aliases, such parameter should be
4080 turned into one passed by value. Return NULL otherwise. */
4081
4082 static struct access *
4083 unmodified_by_ref_scalar_representative (tree parm)
4084 {
4085 int i, access_count;
4086 struct access *repr;
4087 vec<access_p> *access_vec;
4088
4089 access_vec = get_base_access_vector (parm);
4090 gcc_assert (access_vec);
4091 repr = (*access_vec)[0];
4092 if (repr->write)
4093 return NULL;
4094 repr->group_representative = repr;
4095
4096 access_count = access_vec->length ();
4097 for (i = 1; i < access_count; i++)
4098 {
4099 struct access *access = (*access_vec)[i];
4100 if (access->write)
4101 return NULL;
4102 access->group_representative = repr;
4103 access->next_sibling = repr->next_sibling;
4104 repr->next_sibling = access;
4105 }
4106
4107 repr->grp_read = 1;
4108 repr->grp_scalar_ptr = 1;
4109 return repr;
4110 }
4111
4112 /* Return true iff this ACCESS precludes IPA-SRA of the parameter it is
4113 associated with. REQ_ALIGN is the minimum required alignment. */
4114
4115 static bool
4116 access_precludes_ipa_sra_p (struct access *access, unsigned int req_align)
4117 {
4118 unsigned int exp_align;
4119 /* Avoid issues such as the second simple testcase in PR 42025. The problem
4120 is incompatible assign in a call statement (and possibly even in asm
4121 statements). This can be relaxed by using a new temporary but only for
4122 non-TREE_ADDRESSABLE types and is probably not worth the complexity. (In
4123 intraprocedural SRA we deal with this by keeping the old aggregate around,
4124 something we cannot do in IPA-SRA.) */
4125 if (access->write
4126 && (is_gimple_call (access->stmt)
4127 || gimple_code (access->stmt) == GIMPLE_ASM))
4128 return true;
4129
4130 exp_align = get_object_alignment (access->expr);
4131 if (exp_align < req_align)
4132 return true;
4133
4134 return false;
4135 }
4136
4137
4138 /* Sort collected accesses for parameter PARM, identify representatives for
4139 each accessed region and link them together. Return NULL if there are
4140 different but overlapping accesses, return the special ptr value meaning
4141 there are no accesses for this parameter if that is the case and return the
4142 first representative otherwise. Set *RO_GRP if there is a group of accesses
4143 with only read (i.e. no write) accesses. */
4144
4145 static struct access *
4146 splice_param_accesses (tree parm, bool *ro_grp)
4147 {
4148 int i, j, access_count, group_count;
4149 int agg_size, total_size = 0;
4150 struct access *access, *res, **prev_acc_ptr = &res;
4151 vec<access_p> *access_vec;
4152
4153 access_vec = get_base_access_vector (parm);
4154 if (!access_vec)
4155 return &no_accesses_representant;
4156 access_count = access_vec->length ();
4157
4158 access_vec->qsort (compare_access_positions);
4159
4160 i = 0;
4161 total_size = 0;
4162 group_count = 0;
4163 while (i < access_count)
4164 {
4165 bool modification;
4166 tree a1_alias_type;
4167 access = (*access_vec)[i];
4168 modification = access->write;
4169 if (access_precludes_ipa_sra_p (access, TYPE_ALIGN (access->type)))
4170 return NULL;
4171 a1_alias_type = reference_alias_ptr_type (access->expr);
4172
4173 /* Access is about to become group representative unless we find some
4174 nasty overlap which would preclude us from breaking this parameter
4175 apart. */
4176
4177 j = i + 1;
4178 while (j < access_count)
4179 {
4180 struct access *ac2 = (*access_vec)[j];
4181 if (ac2->offset != access->offset)
4182 {
4183 /* All or nothing law for parameters. */
4184 if (access->offset + access->size > ac2->offset)
4185 return NULL;
4186 else
4187 break;
4188 }
4189 else if (ac2->size != access->size)
4190 return NULL;
4191
4192 if (access_precludes_ipa_sra_p (ac2, TYPE_ALIGN (access->type))
4193 || (ac2->type != access->type
4194 && (TREE_ADDRESSABLE (ac2->type)
4195 || TREE_ADDRESSABLE (access->type)))
4196 || (reference_alias_ptr_type (ac2->expr) != a1_alias_type))
4197 return NULL;
4198
4199 modification |= ac2->write;
4200 ac2->group_representative = access;
4201 ac2->next_sibling = access->next_sibling;
4202 access->next_sibling = ac2;
4203 j++;
4204 }
4205
4206 group_count++;
4207 access->grp_maybe_modified = modification;
4208 if (!modification)
4209 *ro_grp = true;
4210 *prev_acc_ptr = access;
4211 prev_acc_ptr = &access->next_grp;
4212 total_size += access->size;
4213 i = j;
4214 }
4215
4216 if (POINTER_TYPE_P (TREE_TYPE (parm)))
4217 agg_size = tree_to_uhwi (TYPE_SIZE (TREE_TYPE (TREE_TYPE (parm))));
4218 else
4219 agg_size = tree_to_uhwi (TYPE_SIZE (TREE_TYPE (parm)));
4220 if (total_size >= agg_size)
4221 return NULL;
4222
4223 gcc_assert (group_count > 0);
4224 return res;
4225 }
4226
4227 /* Decide whether parameters with representative accesses given by REPR should
4228 be reduced into components. */
4229
4230 static int
4231 decide_one_param_reduction (struct access *repr)
4232 {
4233 int total_size, cur_parm_size, agg_size, new_param_count, parm_size_limit;
4234 bool by_ref;
4235 tree parm;
4236
4237 parm = repr->base;
4238 cur_parm_size = tree_to_uhwi (TYPE_SIZE (TREE_TYPE (parm)));
4239 gcc_assert (cur_parm_size > 0);
4240
4241 if (POINTER_TYPE_P (TREE_TYPE (parm)))
4242 {
4243 by_ref = true;
4244 agg_size = tree_to_uhwi (TYPE_SIZE (TREE_TYPE (TREE_TYPE (parm))));
4245 }
4246 else
4247 {
4248 by_ref = false;
4249 agg_size = cur_parm_size;
4250 }
4251
4252 if (dump_file)
4253 {
4254 struct access *acc;
4255 fprintf (dump_file, "Evaluating PARAM group sizes for ");
4256 print_generic_expr (dump_file, parm, 0);
4257 fprintf (dump_file, " (UID: %u): \n", DECL_UID (parm));
4258 for (acc = repr; acc; acc = acc->next_grp)
4259 dump_access (dump_file, acc, true);
4260 }
4261
4262 total_size = 0;
4263 new_param_count = 0;
4264
4265 for (; repr; repr = repr->next_grp)
4266 {
4267 gcc_assert (parm == repr->base);
4268
4269 /* Taking the address of a non-addressable field is verboten. */
4270 if (by_ref && repr->non_addressable)
4271 return 0;
4272
4273 /* Do not decompose a non-BLKmode param in a way that would
4274 create BLKmode params. Especially for by-reference passing
4275 (thus, pointer-type param) this is hardly worthwhile. */
4276 if (DECL_MODE (parm) != BLKmode
4277 && TYPE_MODE (repr->type) == BLKmode)
4278 return 0;
4279
4280 if (!by_ref || (!repr->grp_maybe_modified
4281 && !repr->grp_not_necessarilly_dereferenced))
4282 total_size += repr->size;
4283 else
4284 total_size += cur_parm_size;
4285
4286 new_param_count++;
4287 }
4288
4289 gcc_assert (new_param_count > 0);
4290
4291 if (optimize_function_for_size_p (cfun))
4292 parm_size_limit = cur_parm_size;
4293 else
4294 parm_size_limit = (PARAM_VALUE (PARAM_IPA_SRA_PTR_GROWTH_FACTOR)
4295 * cur_parm_size);
4296
4297 if (total_size < agg_size
4298 && total_size <= parm_size_limit)
4299 {
4300 if (dump_file)
4301 fprintf (dump_file, " ....will be split into %i components\n",
4302 new_param_count);
4303 return new_param_count;
4304 }
4305 else
4306 return 0;
4307 }
4308
4309 /* The order of the following enums is important, we need to do extra work for
4310 UNUSED_PARAMS, BY_VAL_ACCESSES and UNMODIF_BY_REF_ACCESSES. */
4311 enum ipa_splicing_result { NO_GOOD_ACCESS, UNUSED_PARAMS, BY_VAL_ACCESSES,
4312 MODIF_BY_REF_ACCESSES, UNMODIF_BY_REF_ACCESSES };
4313
4314 /* Identify representatives of all accesses to all candidate parameters for
4315 IPA-SRA. Return result based on what representatives have been found. */
4316
4317 static enum ipa_splicing_result
4318 splice_all_param_accesses (vec<access_p> &representatives)
4319 {
4320 enum ipa_splicing_result result = NO_GOOD_ACCESS;
4321 tree parm;
4322 struct access *repr;
4323
4324 representatives.create (func_param_count);
4325
4326 for (parm = DECL_ARGUMENTS (current_function_decl);
4327 parm;
4328 parm = DECL_CHAIN (parm))
4329 {
4330 if (is_unused_scalar_param (parm))
4331 {
4332 representatives.quick_push (&no_accesses_representant);
4333 if (result == NO_GOOD_ACCESS)
4334 result = UNUSED_PARAMS;
4335 }
4336 else if (POINTER_TYPE_P (TREE_TYPE (parm))
4337 && is_gimple_reg_type (TREE_TYPE (TREE_TYPE (parm)))
4338 && bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
4339 {
4340 repr = unmodified_by_ref_scalar_representative (parm);
4341 representatives.quick_push (repr);
4342 if (repr)
4343 result = UNMODIF_BY_REF_ACCESSES;
4344 }
4345 else if (bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
4346 {
4347 bool ro_grp = false;
4348 repr = splice_param_accesses (parm, &ro_grp);
4349 representatives.quick_push (repr);
4350
4351 if (repr && !no_accesses_p (repr))
4352 {
4353 if (POINTER_TYPE_P (TREE_TYPE (parm)))
4354 {
4355 if (ro_grp)
4356 result = UNMODIF_BY_REF_ACCESSES;
4357 else if (result < MODIF_BY_REF_ACCESSES)
4358 result = MODIF_BY_REF_ACCESSES;
4359 }
4360 else if (result < BY_VAL_ACCESSES)
4361 result = BY_VAL_ACCESSES;
4362 }
4363 else if (no_accesses_p (repr) && (result == NO_GOOD_ACCESS))
4364 result = UNUSED_PARAMS;
4365 }
4366 else
4367 representatives.quick_push (NULL);
4368 }
4369
4370 if (result == NO_GOOD_ACCESS)
4371 {
4372 representatives.release ();
4373 return NO_GOOD_ACCESS;
4374 }
4375
4376 return result;
4377 }
4378
4379 /* Return the index of BASE in PARMS. Abort if it is not found. */
4380
4381 static inline int
4382 get_param_index (tree base, vec<tree> parms)
4383 {
4384 int i, len;
4385
4386 len = parms.length ();
4387 for (i = 0; i < len; i++)
4388 if (parms[i] == base)
4389 return i;
4390 gcc_unreachable ();
4391 }
4392
4393 /* Convert the decisions made at the representative level into compact
4394 parameter adjustments. REPRESENTATIVES are pointers to first
4395 representatives of each param accesses, ADJUSTMENTS_COUNT is the expected
4396 final number of adjustments. */
4397
4398 static ipa_parm_adjustment_vec
4399 turn_representatives_into_adjustments (vec<access_p> representatives,
4400 int adjustments_count)
4401 {
4402 vec<tree> parms;
4403 ipa_parm_adjustment_vec adjustments;
4404 tree parm;
4405 int i;
4406
4407 gcc_assert (adjustments_count > 0);
4408 parms = ipa_get_vector_of_formal_parms (current_function_decl);
4409 adjustments.create (adjustments_count);
4410 parm = DECL_ARGUMENTS (current_function_decl);
4411 for (i = 0; i < func_param_count; i++, parm = DECL_CHAIN (parm))
4412 {
4413 struct access *repr = representatives[i];
4414
4415 if (!repr || no_accesses_p (repr))
4416 {
4417 struct ipa_parm_adjustment adj;
4418
4419 memset (&adj, 0, sizeof (adj));
4420 adj.base_index = get_param_index (parm, parms);
4421 adj.base = parm;
4422 if (!repr)
4423 adj.op = IPA_PARM_OP_COPY;
4424 else
4425 adj.op = IPA_PARM_OP_REMOVE;
4426 adj.arg_prefix = "ISRA";
4427 adjustments.quick_push (adj);
4428 }
4429 else
4430 {
4431 struct ipa_parm_adjustment adj;
4432 int index = get_param_index (parm, parms);
4433
4434 for (; repr; repr = repr->next_grp)
4435 {
4436 memset (&adj, 0, sizeof (adj));
4437 gcc_assert (repr->base == parm);
4438 adj.base_index = index;
4439 adj.base = repr->base;
4440 adj.type = repr->type;
4441 adj.alias_ptr_type = reference_alias_ptr_type (repr->expr);
4442 adj.offset = repr->offset;
4443 adj.by_ref = (POINTER_TYPE_P (TREE_TYPE (repr->base))
4444 && (repr->grp_maybe_modified
4445 || repr->grp_not_necessarilly_dereferenced));
4446 adj.arg_prefix = "ISRA";
4447 adjustments.quick_push (adj);
4448 }
4449 }
4450 }
4451 parms.release ();
4452 return adjustments;
4453 }
4454
4455 /* Analyze the collected accesses and produce a plan what to do with the
4456 parameters in the form of adjustments, NULL meaning nothing. */
4457
4458 static ipa_parm_adjustment_vec
4459 analyze_all_param_acesses (void)
4460 {
4461 enum ipa_splicing_result repr_state;
4462 bool proceed = false;
4463 int i, adjustments_count = 0;
4464 vec<access_p> representatives;
4465 ipa_parm_adjustment_vec adjustments;
4466
4467 repr_state = splice_all_param_accesses (representatives);
4468 if (repr_state == NO_GOOD_ACCESS)
4469 return ipa_parm_adjustment_vec ();
4470
4471 /* If there are any parameters passed by reference which are not modified
4472 directly, we need to check whether they can be modified indirectly. */
4473 if (repr_state == UNMODIF_BY_REF_ACCESSES)
4474 {
4475 analyze_caller_dereference_legality (representatives);
4476 analyze_modified_params (representatives);
4477 }
4478
4479 for (i = 0; i < func_param_count; i++)
4480 {
4481 struct access *repr = representatives[i];
4482
4483 if (repr && !no_accesses_p (repr))
4484 {
4485 if (repr->grp_scalar_ptr)
4486 {
4487 adjustments_count++;
4488 if (repr->grp_not_necessarilly_dereferenced
4489 || repr->grp_maybe_modified)
4490 representatives[i] = NULL;
4491 else
4492 {
4493 proceed = true;
4494 sra_stats.scalar_by_ref_to_by_val++;
4495 }
4496 }
4497 else
4498 {
4499 int new_components = decide_one_param_reduction (repr);
4500
4501 if (new_components == 0)
4502 {
4503 representatives[i] = NULL;
4504 adjustments_count++;
4505 }
4506 else
4507 {
4508 adjustments_count += new_components;
4509 sra_stats.aggregate_params_reduced++;
4510 sra_stats.param_reductions_created += new_components;
4511 proceed = true;
4512 }
4513 }
4514 }
4515 else
4516 {
4517 if (no_accesses_p (repr))
4518 {
4519 proceed = true;
4520 sra_stats.deleted_unused_parameters++;
4521 }
4522 adjustments_count++;
4523 }
4524 }
4525
4526 if (!proceed && dump_file)
4527 fprintf (dump_file, "NOT proceeding to change params.\n");
4528
4529 if (proceed)
4530 adjustments = turn_representatives_into_adjustments (representatives,
4531 adjustments_count);
4532 else
4533 adjustments = ipa_parm_adjustment_vec ();
4534
4535 representatives.release ();
4536 return adjustments;
4537 }
4538
4539 /* If a parameter replacement identified by ADJ does not yet exist in the form
4540 of declaration, create it and record it, otherwise return the previously
4541 created one. */
4542
4543 static tree
4544 get_replaced_param_substitute (struct ipa_parm_adjustment *adj)
4545 {
4546 tree repl;
4547 if (!adj->new_ssa_base)
4548 {
4549 char *pretty_name = make_fancy_name (adj->base);
4550
4551 repl = create_tmp_reg (TREE_TYPE (adj->base), "ISR");
4552 DECL_NAME (repl) = get_identifier (pretty_name);
4553 obstack_free (&name_obstack, pretty_name);
4554
4555 adj->new_ssa_base = repl;
4556 }
4557 else
4558 repl = adj->new_ssa_base;
4559 return repl;
4560 }
4561
4562 /* Find the first adjustment for a particular parameter BASE in a vector of
4563 ADJUSTMENTS which is not a copy_param. Return NULL if there is no such
4564 adjustment. */
4565
4566 static struct ipa_parm_adjustment *
4567 get_adjustment_for_base (ipa_parm_adjustment_vec adjustments, tree base)
4568 {
4569 int i, len;
4570
4571 len = adjustments.length ();
4572 for (i = 0; i < len; i++)
4573 {
4574 struct ipa_parm_adjustment *adj;
4575
4576 adj = &adjustments[i];
4577 if (adj->op != IPA_PARM_OP_COPY && adj->base == base)
4578 return adj;
4579 }
4580
4581 return NULL;
4582 }
4583
4584 /* If the statement STMT defines an SSA_NAME of a parameter which is to be
4585 removed because its value is not used, replace the SSA_NAME with a one
4586 relating to a created VAR_DECL together all of its uses and return true.
4587 ADJUSTMENTS is a pointer to an adjustments vector. */
4588
4589 static bool
4590 replace_removed_params_ssa_names (gimple stmt,
4591 ipa_parm_adjustment_vec adjustments)
4592 {
4593 struct ipa_parm_adjustment *adj;
4594 tree lhs, decl, repl, name;
4595
4596 if (gimple_code (stmt) == GIMPLE_PHI)
4597 lhs = gimple_phi_result (stmt);
4598 else if (is_gimple_assign (stmt))
4599 lhs = gimple_assign_lhs (stmt);
4600 else if (is_gimple_call (stmt))
4601 lhs = gimple_call_lhs (stmt);
4602 else
4603 gcc_unreachable ();
4604
4605 if (TREE_CODE (lhs) != SSA_NAME)
4606 return false;
4607
4608 decl = SSA_NAME_VAR (lhs);
4609 if (decl == NULL_TREE
4610 || TREE_CODE (decl) != PARM_DECL)
4611 return false;
4612
4613 adj = get_adjustment_for_base (adjustments, decl);
4614 if (!adj)
4615 return false;
4616
4617 repl = get_replaced_param_substitute (adj);
4618 name = make_ssa_name (repl, stmt);
4619
4620 if (dump_file)
4621 {
4622 fprintf (dump_file, "replacing an SSA name of a removed param ");
4623 print_generic_expr (dump_file, lhs, 0);
4624 fprintf (dump_file, " with ");
4625 print_generic_expr (dump_file, name, 0);
4626 fprintf (dump_file, "\n");
4627 }
4628
4629 if (is_gimple_assign (stmt))
4630 gimple_assign_set_lhs (stmt, name);
4631 else if (is_gimple_call (stmt))
4632 gimple_call_set_lhs (stmt, name);
4633 else
4634 gimple_phi_set_result (as_a <gphi *> (stmt), name);
4635
4636 replace_uses_by (lhs, name);
4637 release_ssa_name (lhs);
4638 return true;
4639 }
4640
4641 /* If the statement STMT contains any expressions that need to replaced with a
4642 different one as noted by ADJUSTMENTS, do so. Handle any potential type
4643 incompatibilities (GSI is used to accommodate conversion statements and must
4644 point to the statement). Return true iff the statement was modified. */
4645
4646 static bool
4647 sra_ipa_modify_assign (gimple stmt, gimple_stmt_iterator *gsi,
4648 ipa_parm_adjustment_vec adjustments)
4649 {
4650 tree *lhs_p, *rhs_p;
4651 bool any;
4652
4653 if (!gimple_assign_single_p (stmt))
4654 return false;
4655
4656 rhs_p = gimple_assign_rhs1_ptr (stmt);
4657 lhs_p = gimple_assign_lhs_ptr (stmt);
4658
4659 any = ipa_modify_expr (rhs_p, false, adjustments);
4660 any |= ipa_modify_expr (lhs_p, false, adjustments);
4661 if (any)
4662 {
4663 tree new_rhs = NULL_TREE;
4664
4665 if (!useless_type_conversion_p (TREE_TYPE (*lhs_p), TREE_TYPE (*rhs_p)))
4666 {
4667 if (TREE_CODE (*rhs_p) == CONSTRUCTOR)
4668 {
4669 /* V_C_Es of constructors can cause trouble (PR 42714). */
4670 if (is_gimple_reg_type (TREE_TYPE (*lhs_p)))
4671 *rhs_p = build_zero_cst (TREE_TYPE (*lhs_p));
4672 else
4673 *rhs_p = build_constructor (TREE_TYPE (*lhs_p),
4674 NULL);
4675 }
4676 else
4677 new_rhs = fold_build1_loc (gimple_location (stmt),
4678 VIEW_CONVERT_EXPR, TREE_TYPE (*lhs_p),
4679 *rhs_p);
4680 }
4681 else if (REFERENCE_CLASS_P (*rhs_p)
4682 && is_gimple_reg_type (TREE_TYPE (*lhs_p))
4683 && !is_gimple_reg (*lhs_p))
4684 /* This can happen when an assignment in between two single field
4685 structures is turned into an assignment in between two pointers to
4686 scalars (PR 42237). */
4687 new_rhs = *rhs_p;
4688
4689 if (new_rhs)
4690 {
4691 tree tmp = force_gimple_operand_gsi (gsi, new_rhs, true, NULL_TREE,
4692 true, GSI_SAME_STMT);
4693
4694 gimple_assign_set_rhs_from_tree (gsi, tmp);
4695 }
4696
4697 return true;
4698 }
4699
4700 return false;
4701 }
4702
4703 /* Traverse the function body and all modifications as described in
4704 ADJUSTMENTS. Return true iff the CFG has been changed. */
4705
4706 bool
4707 ipa_sra_modify_function_body (ipa_parm_adjustment_vec adjustments)
4708 {
4709 bool cfg_changed = false;
4710 basic_block bb;
4711
4712 FOR_EACH_BB_FN (bb, cfun)
4713 {
4714 gimple_stmt_iterator gsi;
4715
4716 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
4717 replace_removed_params_ssa_names (gsi_stmt (gsi), adjustments);
4718
4719 gsi = gsi_start_bb (bb);
4720 while (!gsi_end_p (gsi))
4721 {
4722 gimple stmt = gsi_stmt (gsi);
4723 bool modified = false;
4724 tree *t;
4725 unsigned i;
4726
4727 switch (gimple_code (stmt))
4728 {
4729 case GIMPLE_RETURN:
4730 t = gimple_return_retval_ptr (as_a <greturn *> (stmt));
4731 if (*t != NULL_TREE)
4732 modified |= ipa_modify_expr (t, true, adjustments);
4733 break;
4734
4735 case GIMPLE_ASSIGN:
4736 modified |= sra_ipa_modify_assign (stmt, &gsi, adjustments);
4737 modified |= replace_removed_params_ssa_names (stmt, adjustments);
4738 break;
4739
4740 case GIMPLE_CALL:
4741 /* Operands must be processed before the lhs. */
4742 for (i = 0; i < gimple_call_num_args (stmt); i++)
4743 {
4744 t = gimple_call_arg_ptr (stmt, i);
4745 modified |= ipa_modify_expr (t, true, adjustments);
4746 }
4747
4748 if (gimple_call_lhs (stmt))
4749 {
4750 t = gimple_call_lhs_ptr (stmt);
4751 modified |= ipa_modify_expr (t, false, adjustments);
4752 modified |= replace_removed_params_ssa_names (stmt,
4753 adjustments);
4754 }
4755 break;
4756
4757 case GIMPLE_ASM:
4758 {
4759 gasm *asm_stmt = as_a <gasm *> (stmt);
4760 for (i = 0; i < gimple_asm_ninputs (asm_stmt); i++)
4761 {
4762 t = &TREE_VALUE (gimple_asm_input_op (asm_stmt, i));
4763 modified |= ipa_modify_expr (t, true, adjustments);
4764 }
4765 for (i = 0; i < gimple_asm_noutputs (asm_stmt); i++)
4766 {
4767 t = &TREE_VALUE (gimple_asm_output_op (asm_stmt, i));
4768 modified |= ipa_modify_expr (t, false, adjustments);
4769 }
4770 }
4771 break;
4772
4773 default:
4774 break;
4775 }
4776
4777 if (modified)
4778 {
4779 update_stmt (stmt);
4780 if (maybe_clean_eh_stmt (stmt)
4781 && gimple_purge_dead_eh_edges (gimple_bb (stmt)))
4782 cfg_changed = true;
4783 }
4784 gsi_next (&gsi);
4785 }
4786 }
4787
4788 return cfg_changed;
4789 }
4790
4791 /* Call gimple_debug_bind_reset_value on all debug statements describing
4792 gimple register parameters that are being removed or replaced. */
4793
4794 static void
4795 sra_ipa_reset_debug_stmts (ipa_parm_adjustment_vec adjustments)
4796 {
4797 int i, len;
4798 gimple_stmt_iterator *gsip = NULL, gsi;
4799
4800 if (MAY_HAVE_DEBUG_STMTS && single_succ_p (ENTRY_BLOCK_PTR_FOR_FN (cfun)))
4801 {
4802 gsi = gsi_after_labels (single_succ (ENTRY_BLOCK_PTR_FOR_FN (cfun)));
4803 gsip = &gsi;
4804 }
4805 len = adjustments.length ();
4806 for (i = 0; i < len; i++)
4807 {
4808 struct ipa_parm_adjustment *adj;
4809 imm_use_iterator ui;
4810 gimple stmt;
4811 gdebug *def_temp;
4812 tree name, vexpr, copy = NULL_TREE;
4813 use_operand_p use_p;
4814
4815 adj = &adjustments[i];
4816 if (adj->op == IPA_PARM_OP_COPY || !is_gimple_reg (adj->base))
4817 continue;
4818 name = ssa_default_def (cfun, adj->base);
4819 vexpr = NULL;
4820 if (name)
4821 FOR_EACH_IMM_USE_STMT (stmt, ui, name)
4822 {
4823 if (gimple_clobber_p (stmt))
4824 {
4825 gimple_stmt_iterator cgsi = gsi_for_stmt (stmt);
4826 unlink_stmt_vdef (stmt);
4827 gsi_remove (&cgsi, true);
4828 release_defs (stmt);
4829 continue;
4830 }
4831 /* All other users must have been removed by
4832 ipa_sra_modify_function_body. */
4833 gcc_assert (is_gimple_debug (stmt));
4834 if (vexpr == NULL && gsip != NULL)
4835 {
4836 gcc_assert (TREE_CODE (adj->base) == PARM_DECL);
4837 vexpr = make_node (DEBUG_EXPR_DECL);
4838 def_temp = gimple_build_debug_source_bind (vexpr, adj->base,
4839 NULL);
4840 DECL_ARTIFICIAL (vexpr) = 1;
4841 TREE_TYPE (vexpr) = TREE_TYPE (name);
4842 DECL_MODE (vexpr) = DECL_MODE (adj->base);
4843 gsi_insert_before (gsip, def_temp, GSI_SAME_STMT);
4844 }
4845 if (vexpr)
4846 {
4847 FOR_EACH_IMM_USE_ON_STMT (use_p, ui)
4848 SET_USE (use_p, vexpr);
4849 }
4850 else
4851 gimple_debug_bind_reset_value (stmt);
4852 update_stmt (stmt);
4853 }
4854 /* Create a VAR_DECL for debug info purposes. */
4855 if (!DECL_IGNORED_P (adj->base))
4856 {
4857 copy = build_decl (DECL_SOURCE_LOCATION (current_function_decl),
4858 VAR_DECL, DECL_NAME (adj->base),
4859 TREE_TYPE (adj->base));
4860 if (DECL_PT_UID_SET_P (adj->base))
4861 SET_DECL_PT_UID (copy, DECL_PT_UID (adj->base));
4862 TREE_ADDRESSABLE (copy) = TREE_ADDRESSABLE (adj->base);
4863 TREE_READONLY (copy) = TREE_READONLY (adj->base);
4864 TREE_THIS_VOLATILE (copy) = TREE_THIS_VOLATILE (adj->base);
4865 DECL_GIMPLE_REG_P (copy) = DECL_GIMPLE_REG_P (adj->base);
4866 DECL_ARTIFICIAL (copy) = DECL_ARTIFICIAL (adj->base);
4867 DECL_IGNORED_P (copy) = DECL_IGNORED_P (adj->base);
4868 DECL_ABSTRACT_ORIGIN (copy) = DECL_ORIGIN (adj->base);
4869 DECL_SEEN_IN_BIND_EXPR_P (copy) = 1;
4870 SET_DECL_RTL (copy, 0);
4871 TREE_USED (copy) = 1;
4872 DECL_CONTEXT (copy) = current_function_decl;
4873 add_local_decl (cfun, copy);
4874 DECL_CHAIN (copy) =
4875 BLOCK_VARS (DECL_INITIAL (current_function_decl));
4876 BLOCK_VARS (DECL_INITIAL (current_function_decl)) = copy;
4877 }
4878 if (gsip != NULL && copy && target_for_debug_bind (adj->base))
4879 {
4880 gcc_assert (TREE_CODE (adj->base) == PARM_DECL);
4881 if (vexpr)
4882 def_temp = gimple_build_debug_bind (copy, vexpr, NULL);
4883 else
4884 def_temp = gimple_build_debug_source_bind (copy, adj->base,
4885 NULL);
4886 gsi_insert_before (gsip, def_temp, GSI_SAME_STMT);
4887 }
4888 }
4889 }
4890
4891 /* Return false if all callers have at least as many actual arguments as there
4892 are formal parameters in the current function and that their types
4893 match. */
4894
4895 static bool
4896 some_callers_have_mismatched_arguments_p (struct cgraph_node *node,
4897 void *data ATTRIBUTE_UNUSED)
4898 {
4899 struct cgraph_edge *cs;
4900 for (cs = node->callers; cs; cs = cs->next_caller)
4901 if (!cs->call_stmt || !callsite_arguments_match_p (cs->call_stmt))
4902 return true;
4903
4904 return false;
4905 }
4906
4907 /* Return false if all callers have vuse attached to a call statement. */
4908
4909 static bool
4910 some_callers_have_no_vuse_p (struct cgraph_node *node,
4911 void *data ATTRIBUTE_UNUSED)
4912 {
4913 struct cgraph_edge *cs;
4914 for (cs = node->callers; cs; cs = cs->next_caller)
4915 if (!cs->call_stmt || !gimple_vuse (cs->call_stmt))
4916 return true;
4917
4918 return false;
4919 }
4920
4921 /* Convert all callers of NODE. */
4922
4923 static bool
4924 convert_callers_for_node (struct cgraph_node *node,
4925 void *data)
4926 {
4927 ipa_parm_adjustment_vec *adjustments = (ipa_parm_adjustment_vec *) data;
4928 bitmap recomputed_callers = BITMAP_ALLOC (NULL);
4929 struct cgraph_edge *cs;
4930
4931 for (cs = node->callers; cs; cs = cs->next_caller)
4932 {
4933 push_cfun (DECL_STRUCT_FUNCTION (cs->caller->decl));
4934
4935 if (dump_file)
4936 fprintf (dump_file, "Adjusting call %s/%i -> %s/%i\n",
4937 xstrdup (cs->caller->name ()),
4938 cs->caller->order,
4939 xstrdup (cs->callee->name ()),
4940 cs->callee->order);
4941
4942 ipa_modify_call_arguments (cs, cs->call_stmt, *adjustments);
4943
4944 pop_cfun ();
4945 }
4946
4947 for (cs = node->callers; cs; cs = cs->next_caller)
4948 if (bitmap_set_bit (recomputed_callers, cs->caller->uid)
4949 && gimple_in_ssa_p (DECL_STRUCT_FUNCTION (cs->caller->decl)))
4950 compute_inline_parameters (cs->caller, true);
4951 BITMAP_FREE (recomputed_callers);
4952
4953 return true;
4954 }
4955
4956 /* Convert all callers of NODE to pass parameters as given in ADJUSTMENTS. */
4957
4958 static void
4959 convert_callers (struct cgraph_node *node, tree old_decl,
4960 ipa_parm_adjustment_vec adjustments)
4961 {
4962 basic_block this_block;
4963
4964 node->call_for_symbol_and_aliases (convert_callers_for_node,
4965 &adjustments, false);
4966
4967 if (!encountered_recursive_call)
4968 return;
4969
4970 FOR_EACH_BB_FN (this_block, cfun)
4971 {
4972 gimple_stmt_iterator gsi;
4973
4974 for (gsi = gsi_start_bb (this_block); !gsi_end_p (gsi); gsi_next (&gsi))
4975 {
4976 gcall *stmt;
4977 tree call_fndecl;
4978 stmt = dyn_cast <gcall *> (gsi_stmt (gsi));
4979 if (!stmt)
4980 continue;
4981 call_fndecl = gimple_call_fndecl (stmt);
4982 if (call_fndecl == old_decl)
4983 {
4984 if (dump_file)
4985 fprintf (dump_file, "Adjusting recursive call");
4986 gimple_call_set_fndecl (stmt, node->decl);
4987 ipa_modify_call_arguments (NULL, stmt, adjustments);
4988 }
4989 }
4990 }
4991
4992 return;
4993 }
4994
4995 /* Perform all the modification required in IPA-SRA for NODE to have parameters
4996 as given in ADJUSTMENTS. Return true iff the CFG has been changed. */
4997
4998 static bool
4999 modify_function (struct cgraph_node *node, ipa_parm_adjustment_vec adjustments)
5000 {
5001 struct cgraph_node *new_node;
5002 bool cfg_changed;
5003
5004 cgraph_edge::rebuild_edges ();
5005 free_dominance_info (CDI_DOMINATORS);
5006 pop_cfun ();
5007
5008 /* This must be done after rebuilding cgraph edges for node above.
5009 Otherwise any recursive calls to node that are recorded in
5010 redirect_callers will be corrupted. */
5011 vec<cgraph_edge *> redirect_callers = node->collect_callers ();
5012 new_node = node->create_version_clone_with_body (redirect_callers, NULL,
5013 NULL, false, NULL, NULL,
5014 "isra");
5015 redirect_callers.release ();
5016
5017 push_cfun (DECL_STRUCT_FUNCTION (new_node->decl));
5018 ipa_modify_formal_parameters (current_function_decl, adjustments);
5019 cfg_changed = ipa_sra_modify_function_body (adjustments);
5020 sra_ipa_reset_debug_stmts (adjustments);
5021 convert_callers (new_node, node->decl, adjustments);
5022 new_node->make_local ();
5023 return cfg_changed;
5024 }
5025
5026 /* Means of communication between ipa_sra_check_caller and
5027 ipa_sra_preliminary_function_checks. */
5028
5029 struct ipa_sra_check_caller_data
5030 {
5031 bool has_callers;
5032 bool bad_arg_alignment;
5033 bool has_thunk;
5034 };
5035
5036 /* If NODE has a caller, mark that fact in DATA which is pointer to
5037 ipa_sra_check_caller_data. Also check all aggregate arguments in all known
5038 calls if they are unit aligned and if not, set the appropriate flag in DATA
5039 too. */
5040
5041 static bool
5042 ipa_sra_check_caller (struct cgraph_node *node, void *data)
5043 {
5044 if (!node->callers)
5045 return false;
5046
5047 struct ipa_sra_check_caller_data *iscc;
5048 iscc = (struct ipa_sra_check_caller_data *) data;
5049 iscc->has_callers = true;
5050
5051 for (cgraph_edge *cs = node->callers; cs; cs = cs->next_caller)
5052 {
5053 if (cs->caller->thunk.thunk_p)
5054 {
5055 iscc->has_thunk = true;
5056 return true;
5057 }
5058 gimple call_stmt = cs->call_stmt;
5059 unsigned count = gimple_call_num_args (call_stmt);
5060 for (unsigned i = 0; i < count; i++)
5061 {
5062 tree arg = gimple_call_arg (call_stmt, i);
5063 if (is_gimple_reg (arg))
5064 continue;
5065
5066 tree offset;
5067 HOST_WIDE_INT bitsize, bitpos;
5068 machine_mode mode;
5069 int unsignedp, volatilep = 0;
5070 get_inner_reference (arg, &bitsize, &bitpos, &offset, &mode,
5071 &unsignedp, &volatilep, false);
5072 if (bitpos % BITS_PER_UNIT)
5073 {
5074 iscc->bad_arg_alignment = true;
5075 return true;
5076 }
5077 }
5078 }
5079
5080 return false;
5081 }
5082
5083 /* Return false the function is apparently unsuitable for IPA-SRA based on it's
5084 attributes, return true otherwise. NODE is the cgraph node of the current
5085 function. */
5086
5087 static bool
5088 ipa_sra_preliminary_function_checks (struct cgraph_node *node)
5089 {
5090 if (!node->can_be_local_p ())
5091 {
5092 if (dump_file)
5093 fprintf (dump_file, "Function not local to this compilation unit.\n");
5094 return false;
5095 }
5096
5097 if (!node->local.can_change_signature)
5098 {
5099 if (dump_file)
5100 fprintf (dump_file, "Function can not change signature.\n");
5101 return false;
5102 }
5103
5104 if (!tree_versionable_function_p (node->decl))
5105 {
5106 if (dump_file)
5107 fprintf (dump_file, "Function is not versionable.\n");
5108 return false;
5109 }
5110
5111 if (!opt_for_fn (node->decl, optimize)
5112 || !opt_for_fn (node->decl, flag_ipa_sra))
5113 {
5114 if (dump_file)
5115 fprintf (dump_file, "Function not optimized.\n");
5116 return false;
5117 }
5118
5119 if (DECL_VIRTUAL_P (current_function_decl))
5120 {
5121 if (dump_file)
5122 fprintf (dump_file, "Function is a virtual method.\n");
5123 return false;
5124 }
5125
5126 if ((DECL_ONE_ONLY (node->decl) || DECL_EXTERNAL (node->decl))
5127 && inline_summaries->get (node)->size >= MAX_INLINE_INSNS_AUTO)
5128 {
5129 if (dump_file)
5130 fprintf (dump_file, "Function too big to be made truly local.\n");
5131 return false;
5132 }
5133
5134 if (cfun->stdarg)
5135 {
5136 if (dump_file)
5137 fprintf (dump_file, "Function uses stdarg. \n");
5138 return false;
5139 }
5140
5141 if (TYPE_ATTRIBUTES (TREE_TYPE (node->decl)))
5142 return false;
5143
5144 if (DECL_DISREGARD_INLINE_LIMITS (node->decl))
5145 {
5146 if (dump_file)
5147 fprintf (dump_file, "Always inline function will be inlined "
5148 "anyway. \n");
5149 return false;
5150 }
5151
5152 struct ipa_sra_check_caller_data iscc;
5153 memset (&iscc, 0, sizeof(iscc));
5154 node->call_for_symbol_and_aliases (ipa_sra_check_caller, &iscc, true);
5155 if (!iscc.has_callers)
5156 {
5157 if (dump_file)
5158 fprintf (dump_file,
5159 "Function has no callers in this compilation unit.\n");
5160 return false;
5161 }
5162
5163 if (iscc.bad_arg_alignment)
5164 {
5165 if (dump_file)
5166 fprintf (dump_file,
5167 "A function call has an argument with non-unit alignment.\n");
5168 return false;
5169 }
5170
5171 if (iscc.has_thunk)
5172 {
5173 if (dump_file)
5174 fprintf (dump_file,
5175 "A has thunk.\n");
5176 return false;
5177 }
5178
5179 return true;
5180 }
5181
5182 /* Perform early interprocedural SRA. */
5183
5184 static unsigned int
5185 ipa_early_sra (void)
5186 {
5187 struct cgraph_node *node = cgraph_node::get (current_function_decl);
5188 ipa_parm_adjustment_vec adjustments;
5189 int ret = 0;
5190
5191 if (!ipa_sra_preliminary_function_checks (node))
5192 return 0;
5193
5194 sra_initialize ();
5195 sra_mode = SRA_MODE_EARLY_IPA;
5196
5197 if (!find_param_candidates ())
5198 {
5199 if (dump_file)
5200 fprintf (dump_file, "Function has no IPA-SRA candidates.\n");
5201 goto simple_out;
5202 }
5203
5204 if (node->call_for_symbol_and_aliases
5205 (some_callers_have_mismatched_arguments_p, NULL, true))
5206 {
5207 if (dump_file)
5208 fprintf (dump_file, "There are callers with insufficient number of "
5209 "arguments or arguments with type mismatches.\n");
5210 goto simple_out;
5211 }
5212
5213 if (node->call_for_symbol_and_aliases
5214 (some_callers_have_no_vuse_p, NULL, true))
5215 {
5216 if (dump_file)
5217 fprintf (dump_file, "There are callers with no VUSE attached "
5218 "to a call stmt.\n");
5219 goto simple_out;
5220 }
5221
5222 bb_dereferences = XCNEWVEC (HOST_WIDE_INT,
5223 func_param_count
5224 * last_basic_block_for_fn (cfun));
5225 final_bbs = BITMAP_ALLOC (NULL);
5226
5227 scan_function ();
5228 if (encountered_apply_args)
5229 {
5230 if (dump_file)
5231 fprintf (dump_file, "Function calls __builtin_apply_args().\n");
5232 goto out;
5233 }
5234
5235 if (encountered_unchangable_recursive_call)
5236 {
5237 if (dump_file)
5238 fprintf (dump_file, "Function calls itself with insufficient "
5239 "number of arguments.\n");
5240 goto out;
5241 }
5242
5243 adjustments = analyze_all_param_acesses ();
5244 if (!adjustments.exists ())
5245 goto out;
5246 if (dump_file)
5247 ipa_dump_param_adjustments (dump_file, adjustments, current_function_decl);
5248
5249 if (modify_function (node, adjustments))
5250 ret = TODO_update_ssa | TODO_cleanup_cfg;
5251 else
5252 ret = TODO_update_ssa;
5253 adjustments.release ();
5254
5255 statistics_counter_event (cfun, "Unused parameters deleted",
5256 sra_stats.deleted_unused_parameters);
5257 statistics_counter_event (cfun, "Scalar parameters converted to by-value",
5258 sra_stats.scalar_by_ref_to_by_val);
5259 statistics_counter_event (cfun, "Aggregate parameters broken up",
5260 sra_stats.aggregate_params_reduced);
5261 statistics_counter_event (cfun, "Aggregate parameter components created",
5262 sra_stats.param_reductions_created);
5263
5264 out:
5265 BITMAP_FREE (final_bbs);
5266 free (bb_dereferences);
5267 simple_out:
5268 sra_deinitialize ();
5269 return ret;
5270 }
5271
5272 namespace {
5273
5274 const pass_data pass_data_early_ipa_sra =
5275 {
5276 GIMPLE_PASS, /* type */
5277 "eipa_sra", /* name */
5278 OPTGROUP_NONE, /* optinfo_flags */
5279 TV_IPA_SRA, /* tv_id */
5280 0, /* properties_required */
5281 0, /* properties_provided */
5282 0, /* properties_destroyed */
5283 0, /* todo_flags_start */
5284 TODO_dump_symtab, /* todo_flags_finish */
5285 };
5286
5287 class pass_early_ipa_sra : public gimple_opt_pass
5288 {
5289 public:
5290 pass_early_ipa_sra (gcc::context *ctxt)
5291 : gimple_opt_pass (pass_data_early_ipa_sra, ctxt)
5292 {}
5293
5294 /* opt_pass methods: */
5295 virtual bool gate (function *) { return flag_ipa_sra && dbg_cnt (eipa_sra); }
5296 virtual unsigned int execute (function *) { return ipa_early_sra (); }
5297
5298 }; // class pass_early_ipa_sra
5299
5300 } // anon namespace
5301
5302 gimple_opt_pass *
5303 make_pass_early_ipa_sra (gcc::context *ctxt)
5304 {
5305 return new pass_early_ipa_sra (ctxt);
5306 }
This page took 0.269187 seconds and 6 git commands to generate.