summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--gcc/ChangeLog9
-rw-r--r--gcc/config/arm/arm.c2
-rw-r--r--gcc/config/rs6000/ppu_intrinsics.h2
-rw-r--r--gcc/config/spu/spu.c2
-rw-r--r--gcc/df-scan.c2
-rw-r--r--gcc/fixed-value.c4
-rw-r--r--gcc/fold-const.c4
-rw-r--r--gcc/ginclude/tgmath.h2
-rw-r--r--gcc/haifa-sched.c4
-rw-r--r--gcc/optabs.c2
-rw-r--r--gcc/recog.c2
-rw-r--r--gcc/sched-deps.c4
-rw-r--r--gcc/sched-int.h2
-rw-r--r--gcc/system.h2
-rw-r--r--gcc/target.h2
-rw-r--r--gcc/tree-ssa-live.c2
-rw-r--r--gcc/tree-vect-transform.c4
-rw-r--r--gcc/tree-vectorizer.c2
-rw-r--r--gcc/tree.def2
19 files changed, 32 insertions, 23 deletions
diff --git a/gcc/ChangeLog b/gcc/ChangeLog
index 60eb7cae1b7..2a7373633e8 100644
--- a/gcc/ChangeLog
+++ b/gcc/ChangeLog
@@ -1,5 +1,14 @@
2007-09-01 Kazu Hirata <kazu@codesourcery.com>
+ * config/arm/arm.c, config/rs6000/ppu_intrinsics.h,
+ config/spu/spu.c, df-scan.c, fixed-value.c, fold-const.c,
+ ginclude/tgmath.h, haifa-sched.c, optabs.c, recog.c,
+ sched-deps.c, sched-int.h, system.h, target.h,
+ tree-ssa-live.c, tree-vect-transform.c, tree-vectorizer.c,
+ tree.def: Fix comment typos.
+
+2007-09-01 Kazu Hirata <kazu@codesourcery.com>
+
* config/m68k/m68k.c (TARGET_DEFAULT_TARGET_FLAGS): Remove.
(override_options): Turn on -mstrict-align on non-ColdFire
targets.
diff --git a/gcc/config/arm/arm.c b/gcc/config/arm/arm.c
index fca424aa577..f36b7264cc9 100644
--- a/gcc/config/arm/arm.c
+++ b/gcc/config/arm/arm.c
@@ -5868,7 +5868,7 @@ vfp3_const_double_index (rtx x)
return -1;
/* Sign, mantissa and exponent are now in the correct form to plug into the
- formulae described in the comment above. */
+ formula described in the comment above. */
return (sign << 7) | ((exponent ^ 3) << 4) | (mantissa - 16);
}
diff --git a/gcc/config/rs6000/ppu_intrinsics.h b/gcc/config/rs6000/ppu_intrinsics.h
index 742c5da134a..b18f42de575 100644
--- a/gcc/config/rs6000/ppu_intrinsics.h
+++ b/gcc/config/rs6000/ppu_intrinsics.h
@@ -188,7 +188,7 @@ typedef int __V4SI __attribute__((vector_size(16)));
#endif /* __powerpc64__ */
#ifdef __powerpc64__
-/* Work around the hadware bug in the current Cell implemention. */
+/* Work around the hardware bug in the current Cell implementation. */
#define __mftb() __extension__ \
({ unsigned long long result; \
__asm__ volatile ("1: mftb %[current_tb]\n" \
diff --git a/gcc/config/spu/spu.c b/gcc/config/spu/spu.c
index 7fcc9429310..00df45b4683 100644
--- a/gcc/config/spu/spu.c
+++ b/gcc/config/spu/spu.c
@@ -5484,7 +5484,7 @@ spu_builtin_vectorization_cost (bool runtime_test)
/* If the branch of the runtime test is taken - i.e. - the vectorized
version is skipped - this incurs a misprediction cost (because the
vectorized version is expected to be the fall-through). So we subtract
- the latency of a mispredicted branch from the costs that are incured
+ the latency of a mispredicted branch from the costs that are incurred
when the vectorized version is executed. */
if (runtime_test)
return -19;
diff --git a/gcc/df-scan.c b/gcc/df-scan.c
index 08875038a5a..6c2c6fad58a 100644
--- a/gcc/df-scan.c
+++ b/gcc/df-scan.c
@@ -3632,7 +3632,7 @@ df_record_entry_block_defs (bitmap entry_block_defs)
}
-/* Update the defs in the entry bolck. */
+/* Update the defs in the entry block. */
void
df_update_entry_block_defs (void)
diff --git a/gcc/fixed-value.c b/gcc/fixed-value.c
index 705f4ca3de7..aca386a6d9c 100644
--- a/gcc/fixed-value.c
+++ b/gcc/fixed-value.c
@@ -553,7 +553,7 @@ do_fixed_divide (FIXED_VALUE_TYPE *f, const FIXED_VALUE_TYPE *a,
&r.low, &r.high, 0);
}
- /* Divide r by pos_b to quo_r. The remanider is in mod. */
+ /* Divide r by pos_b to quo_r. The remainder is in mod. */
div_and_round_double (TRUNC_DIV_EXPR, 1, r.low, r.high, pos_b.low,
pos_b.high, &quo_r.low, &quo_r.high, &mod.low,
&mod.high);
@@ -613,7 +613,7 @@ do_fixed_divide (FIXED_VALUE_TYPE *f, const FIXED_VALUE_TYPE *a,
return overflow_p;
}
-/* Calculate F = A << B if LEFT_P. Otherwies, F = A >> B.
+/* Calculate F = A << B if LEFT_P. Otherwise, F = A >> B.
If SAT_P, saturate the result to the max or the min.
Return true, if !SAT_P and overflow. */
diff --git a/gcc/fold-const.c b/gcc/fold-const.c
index ed05f1fc2c0..7220711539a 100644
--- a/gcc/fold-const.c
+++ b/gcc/fold-const.c
@@ -3537,7 +3537,7 @@ omit_one_operand (tree type, tree result, tree omitted)
{
tree t = fold_convert (type, result);
- /* If the resulting operand is an empty statement, just return the ommited
+ /* If the resulting operand is an empty statement, just return the omitted
statement casted to void. */
if (IS_EMPTY_STMT (t) && TREE_SIDE_EFFECTS (omitted))
return build1 (NOP_EXPR, void_type_node, fold_ignored_result (omitted));
@@ -3555,7 +3555,7 @@ pedantic_omit_one_operand (tree type, tree result, tree omitted)
{
tree t = fold_convert (type, result);
- /* If the resulting operand is an empty statement, just return the ommited
+ /* If the resulting operand is an empty statement, just return the omitted
statement casted to void. */
if (IS_EMPTY_STMT (t) && TREE_SIDE_EFFECTS (omitted))
return build1 (NOP_EXPR, void_type_node, fold_ignored_result (omitted));
diff --git a/gcc/ginclude/tgmath.h b/gcc/ginclude/tgmath.h
index 2327556f1a6..1c0fb37cfff 100644
--- a/gcc/ginclude/tgmath.h
+++ b/gcc/ginclude/tgmath.h
@@ -50,7 +50,7 @@ Boston, MA 02111-1307, USA. */
If any generic parameter is complex, we use a complex version. Otherwise
we use a real version. If the real part of any generic parameter is long
double, we use the long double version. Otherwise if the real part of any
- generic paramter is double or of integer type, we use the double version.
+ generic parameter is double or of integer type, we use the double version.
Otherwise we use the float version. */
#define __tg_cplx(expr) \
diff --git a/gcc/haifa-sched.c b/gcc/haifa-sched.c
index 18a5efe03e2..9d1f8b025f5 100644
--- a/gcc/haifa-sched.c
+++ b/gcc/haifa-sched.c
@@ -2924,7 +2924,7 @@ try_ready (rtx next)
else
{
/* One of the NEXT's dependencies has been resolved.
- Recalcute NEXT's status. */
+ Recalculate NEXT's status. */
*ts &= ~SPECULATIVE & ~HARD_DEP;
@@ -3857,7 +3857,7 @@ create_check_block_twin (rtx insn, bool mutate_p)
DONE_SPEC (insn) = ts & BEGIN_SPEC;
CHECK_SPEC (check) = ts & BEGIN_SPEC;
- /* Luckyness of future speculations solely depends upon initial
+ /* Luckiness of future speculations solely depends upon initial
BEGIN speculation. */
if (ts & BEGIN_DATA)
fs = set_dep_weak (fs, BE_IN_DATA, get_dep_weak (ts, BEGIN_DATA));
diff --git a/gcc/optabs.c b/gcc/optabs.c
index 97319714fa5..fee7c5039cc 100644
--- a/gcc/optabs.c
+++ b/gcc/optabs.c
@@ -1265,7 +1265,7 @@ shift_optab_p (optab binoptab)
}
}
-/* Return true if BINOPTAB implements a commutatative binary operation. */
+/* Return true if BINOPTAB implements a commutative binary operation. */
static bool
commutative_optab_p (optab binoptab)
diff --git a/gcc/recog.c b/gcc/recog.c
index 5692f4ef7e2..7ee2f97d3fe 100644
--- a/gcc/recog.c
+++ b/gcc/recog.c
@@ -443,7 +443,7 @@ confirm_change_group (void)
if (changes[i].unshare)
*changes[i].loc = copy_rtx (*changes[i].loc);
- /* Avoid unnecesary rescaning when multiple changes to same instruction
+ /* Avoid unnecesary rescanning when multiple changes to same instruction
are made. */
if (object)
{
diff --git a/gcc/sched-deps.c b/gcc/sched-deps.c
index 433394956ab..cdd055b646f 100644
--- a/gcc/sched-deps.c
+++ b/gcc/sched-deps.c
@@ -359,7 +359,7 @@ free_deps_list (deps_list_t l)
}
/* Return true if there is no dep_nodes and deps_lists out there.
- After the region is scheduled all the depedency nodes and lists
+ After the region is scheduled all the dependency nodes and lists
should [generally] be returned to pool. */
bool
deps_pools_are_empty_p (void)
@@ -648,7 +648,7 @@ sd_finish_insn (rtx insn)
/* Find a dependency between producer PRO and consumer CON.
Search through resolved dependency lists if RESOLVED_P is true.
If no such dependency is found return NULL,
- overwise return the dependency and initialize SD_IT_PTR [if it is nonnull]
+ otherwise return the dependency and initialize SD_IT_PTR [if it is nonnull]
with an iterator pointing to it. */
static dep_t
sd_find_dep_between_no_cache (rtx pro, rtx con, bool resolved_p,
diff --git a/gcc/sched-int.h b/gcc/sched-int.h
index d23fc82d807..da31ec9d8fb 100644
--- a/gcc/sched-int.h
+++ b/gcc/sched-int.h
@@ -897,7 +897,7 @@ struct _sd_iterator
simply a pointer to the next element to allow easy deletion from the
list. When a dep is being removed from the list the iterator
will automatically advance because the value in *linkp will start
- reffering to the next element. */
+ referring to the next element. */
dep_link_t *linkp;
/* True if the current list is a resolved one. */
diff --git a/gcc/system.h b/gcc/system.h
index 84ddf4953ab..c90ded12148 100644
--- a/gcc/system.h
+++ b/gcc/system.h
@@ -789,7 +789,7 @@ extern void fancy_abort (const char *, int, const char *) ATTRIBUTE_NORETURN;
#define CONST_CAST(X) ((void*)(X))
#endif
-/* Acivate -Wcast-qual as a warning (not an error/-Werror). */
+/* Activate -Wcast-qual as a warning (not an error/-Werror). */
#if GCC_VERSION >= 4003
#pragma GCC diagnostic warning "-Wcast-qual"
#endif
diff --git a/gcc/target.h b/gcc/target.h
index e6cb8322e09..5918aed743f 100644
--- a/gcc/target.h
+++ b/gcc/target.h
@@ -306,7 +306,7 @@ struct gcc_target
/* The values of the following two members are pointers to
functions used to simplify the automaton descriptions.
dfa_pre_advance_cycle and dfa_post_advance_cycle are getting called
- immediatelly before and after cycle is advanced. */
+ immediately before and after cycle is advanced. */
void (* dfa_pre_advance_cycle) (void);
void (* dfa_post_advance_cycle) (void);
diff --git a/gcc/tree-ssa-live.c b/gcc/tree-ssa-live.c
index 0fba499fe47..a78dd9c19eb 100644
--- a/gcc/tree-ssa-live.c
+++ b/gcc/tree-ssa-live.c
@@ -477,7 +477,7 @@ remove_unused_scope_block_p (tree scope)
{
next = &TREE_CHAIN (*t);
- /* Debug info of nested function reffers to the block of the
+ /* Debug info of nested function refers to the block of the
function. */
if (TREE_CODE (*t) == FUNCTION_DECL)
unused = false;
diff --git a/gcc/tree-vect-transform.c b/gcc/tree-vect-transform.c
index 00cc8c34218..321d3253419 100644
--- a/gcc/tree-vect-transform.c
+++ b/gcc/tree-vect-transform.c
@@ -4619,7 +4619,7 @@ vect_setup_realignment (tree stmt, block_stmt_iterator *bsi,
The problem arises only if the memory access is in an inner-loop nested
inside LOOP, which is now being vectorized using outer-loop vectorization.
This is the only case when the misalignment of the memory access may not
- remain fixed thtoughout the iterations of the inner-loop (as exaplained in
+ remain fixed throughout the iterations of the inner-loop (as explained in
detail in vect_supportable_dr_alignment). In this case, not only is the
optimized realignment scheme not applicable, but also the misalignment
computation (and generation of the realignment token that is passed to
@@ -6467,7 +6467,7 @@ vect_create_cond_for_align_checks (loop_vec_info loop_vinfo,
DR: The data reference.
VECT_FACTOR: vectorization factor.
- Return an exrpession whose value is the size of segment which will be
+ Return an expression whose value is the size of segment which will be
accessed by DR. */
static tree
diff --git a/gcc/tree-vectorizer.c b/gcc/tree-vectorizer.c
index 2c525eaaa71..11349c2771f 100644
--- a/gcc/tree-vectorizer.c
+++ b/gcc/tree-vectorizer.c
@@ -1701,7 +1701,7 @@ vect_supportable_dr_alignment (struct data_reference *dr)
iterations, it is *not* guaranteed that is will remain the same throughout
the execution of the inner-loop. This is because the inner-loop advances
with the original scalar step (and not in steps of VS). If the inner-loop
- step happens to be a multiple of VS, then the misalignment remaines fixed
+ step happens to be a multiple of VS, then the misalignment remains fixed
and we can use the optimized realignment scheme. For example:
for (i=0; i<N; i++)
diff --git a/gcc/tree.def b/gcc/tree.def
index ea8449b1058..b9c2be1648c 100644
--- a/gcc/tree.def
+++ b/gcc/tree.def
@@ -1017,7 +1017,7 @@ DEFTREECODE (OMP_FOR, "omp_for", tcc_statement, 6)
which of the sections to execute. */
DEFTREECODE (OMP_SECTIONS, "omp_sections", tcc_statement, 3)
-/* This tree immediatelly follows OMP_SECTIONS, and represents the switch
+/* This tree immediately follows OMP_SECTIONS, and represents the switch
used to decide which branch is taken. */
DEFTREECODE (OMP_SECTIONS_SWITCH, "omp_sections_switch", tcc_statement, 0)